Assessment in Learning 1
Assessment in Learning 1
LEARNING GUIDE 3
In Educ 304 – Assessment in Learning 1
First Semester, 2020-2021
Prepared by:
2. Matched the different learning targets and their appropriate assessment techniques;
6. Selected appropriate assessment methods, taking into considerations the different quality
benchmarks; and
SLSU Mission: SLSU will produce S and T leaders and competitive professionals, generate
breakthrough research in S and T –based disciplines, transform and improve the
quality of life in the communities in the service areas, be self-sufficient and
financially variable.
How to use this Learning Guide
This Learning Guide is specially prepared for you, students, enrolled in this course this
semester during this pandemic period when face-to-face classes are not allowed and possible yet.
You should take this module seriously because this employs a self-learning approach and this will
test how diligent, resourceful and interested you are in your dream to become an educator
someday even in this time when we are facing the toughest crisis yet in our lifetime. Needless to
say, consider this module as a challenge and opportunity to prove your place in the Education
Department instead of a heavy burden on your part because your grades largely depend on how
efficient and convincing your answers will be.
Every module is made up of two lessons. Every lesson starts with a brief overview of the
lesson and followed by the intended learning outcomes (ILO) which give you an idea what your
teacher expects from you to learn at the end of the lesson. The lengthy discussions of the topics
immediately follow which hopefully will make you feel like you are in a classroom setting.
After the discussions, an assessment is provided to measure how far have you learned the
discussions. The assessment is broken down to tasks which are assigned to a perfect score each.
You are to comply all the tasks and submit to your teacher because your scores will be the basis of
your midterm and final grades. Every task also has its own specific instruction/s and it is very
important that you follow them strictly. Submit your answers to all tasks in one lesson
altogether. You are reminded to work on your own. Do not just copy paste your answers
from your classmates or friends from the other class, otherwise it may result to substantial
deduction of your scores or worse, total unacceptance of your submitted works. Your
submission can be in a printed (hardcopy) form, or a softcopy through an agreed platform which
will be agreed upon between you and your teacher later, including the deadlines, when online
communication between us has become established.
The learning resources that follow are the possible options where online communication
and submission of answers to assessment (softcopy) maybe done. The module is ended with a list
of references from where the discussions are taken from, including plenty of online sources, which
if you open them, will surely help you understand better and deeper the discussions resulting to an
easier and faster compliance to all tasks in assessment.
There are four modules with two lessons each. A pre-test and post-test are found in first
and last module, respectively, and both have to be answered, too. You have however, to submit your
answers of the pre-test first ahead of the assessment answers to the first lesson while post-test can
be answered only after finishing the last module.
And finally, this module is anchored on the syllabus of this course which will be provided to
you in a separate attachment which will guide you with the flow and continuity of our engagement
through these modules.
Your teacher is very much willing to facilitate your learning and should you have concerns
regarding the modules, please feel free to communicate him anytime of the day, 24/7. Thank you,
good luck and God Bless..
Sir Glenn
Lesson 5. Development of Varied Assessment Tools
Assessment is a systematic process of gathering information about what a student knows, is
able to do, and is learning to do. Assessment information provides the foundation for decision-
making and planning for instruction and learning. Assessment is an integral part of instruction that
enhances, empowers, and celebrates student learning. Using a variety of assessment techniques,
teachers gather information about what students know and are able to do, and provide positive,
supportive feedback to students. They also use this information to diagnose individual needs and to
improve their instructional programs, which in turn helps students learn more effectively.
Discussion
5.1 Planning a Test and Construction of Table of Specifications (ToS)
The important steps in planning for a test are:
Identifying test objectives/lesson outcomes
Deciding on the type of objective test to be prepared
Preparing a Table of Specifications (ToS)
Constructing the draft test items
Try-out and validation
1. Test what you actually taught. Good tests cover the content and objectives the
instructor actually taught—this is called content validity. For example, if you devote a lot
of time to a particular topic in a unit you would use more test items on that topic than on
minor topics. The same is true for objectives. If you devote a lot of time teaching a
particular objective, say the ability to analyze the subject matter, then the test would
include items that assess students’ ability to analyze the material.
2. Improve test items. There are well established rules and guidelines for writing good test
items. Using a few simple rules can improve the quality of your tests.
3. Write items that assess complex learning objectives. A popular criticism of m-c tests
is that they only measure what students remember (recall or recognize] about the course
material. There is nothing wrong with testing for basic knowledge or familiarity with
content, especially if it is important for students to have a working knowledge of
terminology, facts, and principles, etc. But instructors also address more complex
objectives such as understanding, problem solving, application, analysis and evaluation
of the material. Writing m-c items that measure higher-level objectives is challenging but
attainable.
4. Improve overall test construction and administration. Good test items are only one
aspect of a good test. Additional considerations include the clarity of instructions,
organization of items on the test, test length and time limits.
Examples:
Organize blocks of test items according to topics. List all the items for Topic A
followed by all the items for Topic B and so forth. This reduces the cognitive load on
students during the test. If items appear randomly by topic, students must quickly
switch back and forth through the material rather than focusing on a single topic
before moving on to the next.
Provide sufficient time for students to complete the test. Students differ in how
quickly they read, process information and formulate ideas. Pressure to complete
tests quickly can disadvantage slower-responding students and underestimate their
knowledge of the subject.
5. Use feedback to improve test items. Revise test items based on feedback from
colleagues, students and statistical analyses.
Examples:
Feedback from colleagues. Several instructors who teach the same course can
quickly review one another’s test items and develop a shared pool of good items.
Feedback from students. Review test results with students. This is an opportunity to
give feedback to students, and also get feedback from them about test items.
Statistical Item Analysis. If you have tests scored by ITS, you receive an item analysis
that provides item difficulty and item discrimination.
Essentially, a table of specification is a table chart that breaks down the topics that will be on a
test and the amount of test questions or percentage of weight each section will have on the final test
grade. This kind of table chart is usually split into two charts, and each sub topic is numbered under
the main topics that are being covered for the test. This type of table is mainly used by
teachers to help break down their testing outline on a specific subject. Some teachers use this
particular table as their teaching guideline by breaking the table into subjects, the teachers’ main
points much time should be spent on the point, and what assignment or project can be done to help
the student learn the subject.
A prototype table of a periodical test in TLE III-Home Economics subject is shown using the
link below:
https://www.scribd.com/doc/125213887/Table-of-Specs-in-Grade-7-ICT
and for more examples of ToS, please use the link below:
https://depedtambayan.org/new-2017-2018-periodic-test-questions-with-table-of-
specifications/
The ToS guides the teacher in formulating the test. As we can see, the ToS also ensures that
each of the objectives in the hierarchy of educational objectives is well represented in the test. As
such, the resulting test that will be constructed by the teacher will be more or less comprehensive.
Without the ToS, the tendency for the test maker is to focus too much on facts and concepts at the
knowledge level.
For more understanding of the discussion above, the following tutorial videos are
recommended. For this purpose, use the following links below:
https://www.youtube.com/watch?v=6NaPvWNW41s
https://www.youtube.com/watch?v=CrO-dYx2BYU
https://www.youtube.com/watch?v=hrW9p2WziGI
Example: The Philippines gained its independence in 1898 and therefore celebrated its
centennial year in 2000. ______
Obviously, the answer is FALSE because 100 years from 1898 is not 2000 but 1998.
Rule 2: Avoid using the words “always”, “never” “often” and other adverbs that tend to be
either always true or always false.
Statements that use the word “always” are almost always false. A test-wise student can
easily guess his way through a test like these and get high scores even if he does not know
anything about the test.
Rule 3: Avoid long sentences as these tend to be “true”. Keep sentences short.
Example: Tests need to be valid, reliable and useful, although, it would require a great amount
of time and effort to ensure that tests possess these test characteristics. _______
Notice that the statement is true. However, we are also not sure which part of the sentence
is deemed true by the student. It is just fortunate that in this case, all parts of the sentence
are true and hence, the entire sentence is true. The following example illustrates what can
go wrong in long sentences:
Example: Tests need to be valid, reliable and useful since it takes very little amount of time,
money and effort to construct tests with these characteristics.
The first part of the sentence is true but the second part is debatable and may, in fact, be
false. Thus, a “true” response is correct and also, a “false” response is correct.
Rule 4. Avoid trick statements with some minor misleading word or spelling anomaly,
misplaced phrases, etc. A wise student who does not know the subject matter may
detect this strategy and thus get the answer correctly.
The Principal’s name may actually be correct but since the word is misspelled and the entire
sentence takes a different meaning, the answer would be false! This is an example of a tricky
but utterly useless item.
Rule 5: Avoid quoting verbatim from reference materials or textbooks. This practice sends
the wrong signal to the students that it is necessary to memorize the textbook word
for word and thus, acquisition of higher-level thinking skills is not given due
importance.
Rule 7. With true or false questions, avoid a grossly disproportionate number of either true
or false statements or even patterns in the occurrence of true and false statements.
A Multiple Choice Question (MCQ) is one of the most popular assessment methods that
can be used for formative and summative assessments. Here, in response to a question, the
learner is invited to select one correct option from a list of options. It is an effective and efficient
way to assess learning outcomes. Multiple choice test items have several potential advantages:
Versatility, Reliability, and Validity.
Characteristics of MCQs
Multiple choice test items have several potential advantages:
Versatility: Multiple choice tests assess various levels of learning outcomes, from basic
recall to application, analysis, and evaluation. Because learners are choosing from a set of
potential answers, however, there are obvious limits on what can be tested. For example,
they are not an effective way to test learners’ ability to organize thoughts or articulate
explanations or creative ideas.
Validity: Validity is the degree to which a test measures the learning outcomes it claims
to offer. Since learners normally answer a multiple-choice question more quickly than an
essay question, tests based on multiple choice mostly focus on a broader representation
of course material.
Grammatical clues make the answer obvious to learners. We must ensure that the question
stem is free from grammatical clues such as ‘a’ or ‘an’.
Heterogeneous content provides clue to the learner on which answer is correct. Alternatives
should be homogenous and shouldn’t provide a clue. Also, it is important to note that the
alternatives have grammar consistent with the question stem, parallel in form, and similar
in length and language.
The stem should either be a question or a partial sentence. This allows the learner to focus
on answering the question rather than holding the partial sentence in working memory and
sequentially completing it with each option.
The cognitive load on the learner is increased when the stem is constructed with an initial
or blank in- between. So avoid such constructions.
Options that overlap confuse learners and can be easily identified as distractors. Also, if the
overlap includes the intended answer, then there will be more than one option that can be
confused as the right answer.
The function of incorrect options is to serve as distractors. Options that are implausible
don’t serve as distractors and hence, should not be used.
Arranging alternatives logically in an alphabetical or numerical order will help you avoid
cognitive overload and possible confusion.
If you have noticed, correct answers are usually at the second or third positions. Many
instructors have a tendency to bury the answers in the middle, instead of placing them as
the first or last option. This practice, however, is quite known to learners. The best method
is to randomize the answers.
When words such as never, always, and only are included in the distractors to make them
false, they serve as clues to alert learners. Instead, if necessary, you can include them in the
beginning of your question stem rather than using them in the options.
d. Never wipe the wound with d. Wipe the wound with alcohol
alcohol unless it is still bleeding unless it is still bleeding
An option in a multiple-choice question that is noticeably longer or shorter than the others
has a greater chance of being assumed to be the correct answer. Hence, to avoid such
choices without good reason, always keep your options similar in length.
Learners prefer multiple choice questions over other assessment methods because it is
simple and easy to score. By following the abovementioned rules, you can design objective and
reliable multiple-choice questions to evaluate learning.
A B
_________1. Magellan a. First President of the Republic
_________2. Mabini b. National Hero
_________3. Rizal c. Discovered the Philippines
_________4. Lapu-Lapu d. Brain of Katipunan
_________5. Aguinaldo e. The great painter
f. Defended Limasawa island
Normally, column B will contain more items than column A to prevent guessing on the part
of the students. Matching type items, unfortunately, often test lower order thinking skills
(knowledge level) and are unable to test higher order thinking skills such as application and
judgement skills.
A variant of the matching type items is the data sufficiency and comparison type of test
illustrated below:
Example: Write L if the item on the left is greater than the item on the right; O if the item on
the left is less than the item on the right; V if the item on the left equals the item
on the right and E if the relationship cannot be determined.
A B
1. Square root of 9 ______ a. -3
2. Square root of 25 ______ b. 615
3. 36 inches ______ c. 3 meters
4. 4 feet ______ d. 48 inches
5. 1 kilogram ______ e. 1 pound
The data sufficiency test above can, if properly constructed, test higher-order thinking skills.
Each item goes beyond simple recall of facts and, in fact, requires the students to make decisions.
A B
1. The year of the New York’s founding 4 July 1776
2. The capital of the United States. B. George Washington
3. First president of the United States. C. 1653
4. The date the declaration of D. United States dollar
independence was signed.
5. The name of the United States E. Washington
currency.
Correct
A B
1. The largest planet in the Solar System. A. Mercury
2. The planet humans first landed on. B. Neptune
3. The fathest planet from Earth. C. The Moon
4. The planet with an observable ring system D. Jupiter
5. The smallest planet in the Solar System. E. Saturn
A B
1. Google A. Which of the companies on the list derives
its profit primarily from context advertising?
2. Microsoft B. Which of the companies on the list first
produced fonts for an operating system?
3. Apple C. Which of the companies on the list first
proposed using a computer mouse?
D. Which of the companies on the list first
introduced a graphical user interface in their
operating system?
E. Which of the companies on the list first
created a browser-based cloud operating
system?
Classification Questions
These are very similar to matching questions with keylists, the only difference being that the
learners are asked to sort answers from the second column into groups belonging to separate
classes or categories specified in the first column. Classification questions consist of a description of
the task the learner has to perform, the list of elements to be sorted, and the list of categories they
have to be sorted into. Below is an example of a classification question.
Sort the following animals according to the species they belong to:
A B
1. Mammals A. Whale
2. Birds B. Duck
3. Fish C. Dolphin
D. Pelican
E. Salamander
Another useful device for testing lower-order thinking skills is the supply type of tests. Like
the multiple-choice test, the items in this kind of test consist of a stem and a blank where the
students would write the correct answer.
Supply type tests depend heavily on the way that the stems are constructed. These tests
allow for one and only one answer and, hence, often test only the students’ knowledge. It is,
however, possible to construct supply type of tests that will test higher order thinking as the
following example shows:
Example: Write an appropriate synonym for each of the following. Each blank corresponds
to a letter:
Metamorphose: _ _ _ _ _ _
Flourish: _ _ _ _
The appropriate synonym for the first is CHANGE with six (6) letters while the
appropriate synonym for the second is GROW with four (4) letters. Notice that these
questions require not only mere recall of words but also understanding of these words.
Completion items require a word, short phrase, number, or symbol response. They are
sometimes called fill-in-the-blank questions. Sometimes a word bank accompanies a set of short-
answer exercises, especially for younger children or where recognition (not production) of the
word is important and spelling is not. Use completion items especially when recall of a specific
word or term is important for the learning goal. Write effective completion items by following
the guidelines below.
Write each item to tap into the content and kind of thinking in important learning
outcomes. Make sure the alignment between the item and the learning outcome is
strong. For example, consider the learning outcome, “Students will identify three-
dimensional geometric shapes.” An effective item on a geometry test might be, “What
shape is a basketball? ________.” To answer correctly with “sphere,” the student needs to
know what a basketball is (safe to assume for most students), to recall its shape, and to
know that shape is a sphere. The item matches the outcome in both content (i.e., a
sphere is a three-dimensional geometric shape) and level of thinking (i.e., recall).
Effective items will not confound relevant content and thinking skills with other
knowledge or skills not required for the outcome. For example, the question, “What
shape is Spaceship Earth at Epcot? ________” confounds travel experience (or interest in
theme parks) with geometry knowledge. To answer correctly, the student would need
to know what the Spaceship Earth building looked like and to recognize the shape of a
sphere. A student might get this answer wrong not because he or she couldn’t identify a
sphere, but because he or she didn’t know what Spaceship Earth was.
Insert a blank where an answer is expected. If the item is a direct question, a blank
should be placed after the question. If the item is a sentence that leaves out a key word
or idea, a blank that serves as a placeholder for that item should be included and it
should come toward the end of the sentence. When the missing element is placed near
the end of the sentence, students can read through the sentence once and gather
enough context to answer without having to backtrack and read the sentence again. For
example, “What is the most common gas in the earth’s atmosphere? _______” and “The
most common gas in the earth’s atmosphere is ________.” are good items because
students will recognize that they must identify earth’s most common gas on the first
read-through before they come to the blank. In contrast, “________ is the most common
gas in the earth’s atmosphere” is less effective, because it requires students to read to
the end of the sentence to recognize that the blank should be filled with the name of
earth’s most common gas.
Avoid grammatical or other clues. Clues from grammar or from the length of the line
confound other kinds of reasoning with the knowledge or skill the item is intended to
assess. In a set of completion items, keep all blanks the same length so that students
don’t infer that some answers are long and others are short. In addition, avoid grammar
clues within the item. For example, the item “A person who draws the pictures for a book
is called an _______.” suggests that the correct answer begins with a vowel sound.
Grammar clues can be avoided by rephrasing items as direct questions (e.g., “What is a
person who draws the pictures for a book called? _______”) or by providing options that
eliminate the grammatical clue (e.g., “A person who draws the pictures for a book is
called a(n) _______.”
Essays, classified as non-objective tests, allow for the assessment of higher-order thinking
skills. Such tests require students to organize their thoughts on a subject matter in coherent
sentences in order to inform an audience. In essay tests, students are required to write one or more
paragraphs on a specific topic.
Essay questions can be used to measure the attainment of a variety of objectives. Stecklein
(1955) has listed 14 types of abilities that can be measured by essay items:
There are two types of essay items: extended response and restricted response.
An extended response essay item is one that allows for an in-depth sampling of a student’s
knowledge, thinking processes, and problem-solving behavior relative to a specific topic. The open-
ended nature of the task posed by an instruction such as “discuss essay and objective tests” is
challenging to a student. In order to answer this question correctly, the student has to recall specific
information and organize, evaluate, and write an intelligible composition. The potential ambiguity
of an essay task is probably the single most important contributor to unreliability. In addition, the
more extensive the responses required and the fewer questions a teacher may ask would definitely
result to lower content validity of the test.
Examples:
Following are examples of essay questions based on Bloom’s Taxonomy of Cognitive Objectives.
A. Knowledge:
Explain how Egypt came to be called the gift of the Nile.
B. Comprehension
What is meant when the person says, “I had just crossed the bridge”?
C. Application
Give at least three examples of how the law of supply operates in our economy today.
D. Analysis
Explain the causes and effects of the People’s Power Revolution on the political and social
life of the Filipino people.
E. Synthesis
Describe the origin and significance of the celebration of Christmas the world over.
No Needs
Adequate Quality Exemplary
Answer Improvement
cccc
6 pts 8 pts 10 pts
0 pts 4 pts
Content Did not Answers are Answers are Answers are Answers are
10 pts answer partial or not accurate and comprehensive,
question. incomplete. comprehensive complete. accurate and
Key points are or completely Key points complete. Key
not clear. stated. Key are stated ideas are
Question not points are and clearly stated,
adequately addressed, but supported. explained, and
answered. not well well supported.
supported.
Writing Did not Displays over Displays three Displays one Displays no
Conventions answer five errors in to five errors in to three errors in
10 pts question. spelling, spelling, errors in spelling,
punctuation, punctuation, spelling, punctuation,
(Spelling, grammar, and grammar, and punctuation, grammar, and
punctuation, sentence sentence grammar, sentence
grammar, and structure. structure. and sentence structure.
complete structure.
sentences.)
Learning Resources:
1. Module
2. Schoology/Moodle
3. Messenger
References:
1. De Guzman, Estefania S. and Adamos, Joel L. (2015). Assessment in Learning 1. Adriana
Publishing Co., Inc. Quezon City, Manila.
2. Navarro, Rosita L. et. al. (2017). Assessment of Learning 1, Third Edition. Lorimar
Publishing, Inc. Quezon City, Metro Manila.
3. https://www.theclassroom.com/develop-assessment-tools-7771843.html
4. https://www.academia.edu/27158440/
TOOLS_and_TECHNIQUES_FOR_CLASSROOM_ASSESSMENT
5. https://prezi.com/3xsnujo6crs_/planning-an-objective-type-of-test/
6. https://sites.google.com/a/uwlax.edu/teaching-improvement-guide/improvement-
strategies/assessing-learning/b-exams-and-quizzes
7. https://citl.illinois.edu/citl-101/measurement-evaluation/exam-scoring/improving-
your-test-questions
8. https://cehs01.unl.edu/aalbano/intromeasurement/mainch4.html
9. https://www.yourarticlelibrary.com/education/test/top-4-steps-for-constructing-a-
test/64781
10. https://www.academia.edu/33665595/
How_to_create_a_Table_of_Specifications_TOS_in_5_Easy_Steps
11. http://educ5let2012.blogspot.com/2012/08/table-of-specification-essentially.html
12. http://guroako.com/2019/08/04/table-of-specification-template/
13. https://elcomblus.com/rules-in-constructing-true-false-tests/
14. https://www.educationcorner.com/true-false-tests.html
15. http://theelearningcoach.com/elearning_design/rules-for-multiple-choice-questions/
16. https://blog.commlabindia.com/elearning-design/effective-multiple-choice-questions
17. https://www.uww.edu/learn/improving/restiptool/multiple-choice-and-matching
18. https://elcomblus.com/rules-in-constructing-matching-type-and-supply-type-of-tests/
19. https://wps.prenhall.com/chet_stiggins_student_6e/213/54606/13979157.cw/
content/index.html
20. https://elcomblus.com/rules-in-constructing-essay-tests/
21. https://testing.byu.edu/handbooks/WritingEffectiveEssayQuestions.pdf
22. https://www.csbsju.edu/academic-advising/study-skills-guide/essay-exam
23. http://ed103shibboleth.blogspot.com/2011/04/constructing-and-scoring-essay-
tests.html
24. http://educationalevaluationhema.blogspot.com/2011/03/essay-test.html
Lesson 6. Item Analysis and Validity
In “Assessment: The Bridge Between Teaching and Learning,” Dylan William states:
“If our students learned what we taught, we would never need to assess...It is only through
assessment that we can discover whether the instructional activities in which we engaged our
students resulted in the intended learning. Assessment really is the bridge between teaching and
learning.”
Assessment via midterms, tests, quizzes, and exams is the way in which educators gain
insight into student learning; in fact, assessment accounts for well over 50% of a student’s grade in
many higher education courses.
It is also a venue through which educators can address student learning, because exams are
a window into student learning gaps, and consequently, a way to shore up student learning. Exams
should answer a question for educators as much as they offer questions for students:
With which topics are students struggling, and why? One way to increase visibility into
student learning gaps is via item analysis.
Discussion
6.1 What is Item Analysis?
Item analysis is the act of analyzing student responses to individual exam questions with
the intention of evaluating exam quality. It is an important tool to uphold test effectiveness and
fairness.
Item analysis is likely something educators do both consciously and unconsciously on a
regular basis. In fact, grading literally involves studying student responses and the pattern of
student errors, whether to a particular question or particular types of questions. But when the
process is formalized, item analysis becomes a scientific method through which tests can be
improved, and academic integrity upheld.
Item analysis brings to light test quality in the following ways:
Item Difficulty/Difficulty Index
Is the exam question (aka “item”) too easy or too hard? When an item is one that
every student either gets wrong or correct, it decreases an exam’s reliability. If everyone
gets a particular answer correct, there’s less of a way to tell who really understands the
material with deep knowledge. Conversely, if everyone gets a particular answer incorrect,
then there’s no way to differentiate those who’ve learned the material deeply.
An item’s difficulty index is obtained by calculating the p value (p) which is the
proportion of students answering the item correctly.
p = R/T
Item 1: There were 45 students in the class Item 1 has a p value of 0.67. Sixty-seven
who responded to Item 1 and 30 answered percent (67%) got the item right while
it correctly. 33% missed it.
p = 30/45
= 0.67
Item 2. In the same class, only 10 Item 2 has a p value of 0.22. Out of 45
responded correctly to Item 2. only 10 or 22% got the item right while 35
or 78% missed it.
p = 10/45
= 0.22
For Normative -referenced test: Between the two items. Item 2 appears to be a much
more difficult item since less than a fourth of the class was able to respond correctly.
For Criterion-referenced test: The class shows much better performance in Item 1 than
in Item 2. It is still a long way for many to master them.
The p-value ranges from 0.0 to 1.0 which indicates from extremely very difficult as
no one got it correctly to extremely very easy as everyone got it correct. For binary-choice
items, there is a 50% probability of getting the item correct simply by chance. For multiple-
choice items of four alternatives, the chance of obtaining a correct answer by guessing is
only 25%. This is an advantage of using multiple-choice questions over binary-choice items.
The probability of getting a high score by chance is reduced.
For classroom tests, the discrimination index shows if a difference exists between
the performance of those who scored high and those who scored low in an item. As a
general rule, the higher the discrimination index (D), the more marked the magnitude of the
difference is, thus, the more discriminating the item is. The nature of the difference
however, can take different directions.
Another calculation can bring about the same result as (Kubiszyn and Borich, 2010):
D = Ru-Rl
T
As you can see, R/T is actually getting the p-value of an item. So, to get D is to get the
difference between the p-value involving the upper half and the p-value involving the lower
half. So, the formula for discrimination index (D) can also be given as (Popham, 2011):
D = pu-pl
Where pu is the p-value for the upper group (Ru/Tu)
pl is the p-value for the lower group (Rl/Tl)
To obtain the proportions of the upper and the lower groups responding to the item
correctly, the teacher follows these steps:
a. Score the test papers using a key to correction to obtain the total scores of the
students. Maximum score is the total number of objective items.
b. Order the test papers from highest to lowest score.
c. Split the test papers into halves: high group and low group
For a class of 50 or less students, do a 50-50 split. Take the upper half as the
HIGH Group and the lower half as the LOW Group
For a big group of 100 or so, take the upper 25-27% and the lower 25-27%.
Maintain equal numbers of test papers for Upper and Lower group.
d. Obtain the p-value for the Upper group and p-value for the Lower group
e. Get the Discrimination index by getting the difference between the p-values.
For the purposes of evaluating the discriminating power of items, Popham (2011)
offers the guidelines proposed by Ebel & Frisbie (1991) shown in the table below. The
teachers can be guided on how select the satisfactory items and what to do to improve the
rest.
Items with negative discrimination indices, although significantly high, are subject
right away to revision if not deletion. With multiple-choice items, negative D is a forensic
evidence of errors in item writing. It suggests the possibility of:
Wrong key. More knowledgeable students selected a distracter which is the correct
answer but is not the keyed option.
Unclear problem in the stem leading to more than one correct answer.
Ambiguous distracters leading the more informed students be divided in choosing the
attractive options.
Implausible keyed options which more informed students will not choose.
Item Distractors
For multiple-choice exams, distractors play a significant role. Do exam questions
effectively distract test takers from the correct answer? For example, if a multiple-choice
question has four possible answers, are two of the answers obviously incorrect, thereby
rendering the question with a 50/50 percent chance of correct response? When distractors
are ineffective and obviously incorrect as opposed to being more disguised, then they
become ineffective in assessing student knowledge. An effective distractor will attract test
takers with a lower overall score than those with a higher overall score.
Item analysis entails noting the pattern of student errors to various questions in all
the ways stated above. This analysis can provide distinct feedback on exam efficacy and
support exam design.
Item distractor analysis is also helpful in that it can help identify misunderstandings
students have about the material. If the majority of students selected the same incorrect
multiple-choice answer, then that provides insight into student learning needs and
opportunities.
The following links can lead you to some video tutorials that can help you understand
better the discussions above. Note that the first 5 links are a series already sequenced by an
American teacher, while the last 2 links are authored by Filipino teachers.
1. https://www.youtube.com/watch?v=BMI9nihpsPc
2. https://www.youtube.com/watch?v=rQpXL5b8XYw
3. https://www.youtube.com/watch?v=44XCB0I9fk4
4. https://www.youtube.com/watch?v=YLpaLYy8fEo
5. https://www.youtube.com/watch?v=y0n-6rE-Zhs
6. https://www.youtube.com/watch?v=pMR7Eb_xlQE
7. https://www.youtube.com/watch?v=Hg0wRlOlflc
Learning Resources:
4. Module
5. Schoology/Moodle
6. Messenger
References:
1. De Guzman, Estefania S. and Adamos, Joel L. (2015). Assessment in Learning 1. Adriana
Publishing Co., Inc. Quezon City, Manila.
2. Navarro, Rosita L. et. al. (2017). Assessment of Learning 1, Third Edition. Lorimar
Publishing, Inc. Quezon City, Metro Manila.
3. http://www1.udel.edu/educ/gottfredson/451/unit9-guidance.htm
4. https://www.uwosh.edu/testing/faculty-information/test-scoring/score-report-
interpretation/item-analysis-1
5. https://www.uwosh.edu/testing/faculty-information/test-scoring/score-report-
interpretation/item-analysis-1/item-difficulty
6. https://www.uwosh.edu/testing/faculty-information/test-scoring/score-report-
interpretation/item-analysis-1/reliability-validity
7. httpwww.geneseo.edu/sites/default/files/sites/education/p12resources-item-analysis-and-
instruction.pdf
8. https://www.turnitin.com/blog/what-is-item-analysis-and-other-important-exam-design-
principles
9.