A good test should have the following key characteristics:
1. It should be a valid instrument that accurately measures what it is intended to measure as evidenced by various types of validity like content validity.
2. It should be a reliable instrument that consistently measures constructs and yields similar results over time as determined through methods like test-retest reliability.
3. It should be objective by eliminating personal bias and opinions of scorers so that different scorers arrive at the same score.
TEST
ļ¬ a formaland systematic instrument,
usually paper and pencil procedure
designed to assess the quality, ability, skill
or knowledge of the students by giving a
set of question in uniform manner
ļ¬one of the many types of assessment
procedure used to gather information
about the performance of students
ā¢VALIDITY
ļ¬ DEFINITION:
ļ¬ āValidityis the extent to which a test
measures what it claims to measure. It is
vital for a test to be valid in order for the
results to be accurately applied and
interpreted.ā
5.
Other definitions givenby experts
ļ¬Gronlund and Linn (1995)- ā Validity refers
to the appropriateness of the interpretation
made from test scores and other
evaluation results with regard to a
particular use.ā
ļ¬Anne Anastasi (1969) writes ā the validity
of a test concerns what the test measures
and how well it does so.ā
6.
ļ¬Ebel and Frisbie(1991)- ā The term
validity, when applied to a set of test
scores, refers to the consistency
(accuracy) with which the scores measure
a particular cognitive ability of interest.ā
ļ¬C.V. Good (1973)- in the dictionary of
education defines validity as the ā extent to
which a test or other measuring instrument
fulfills the purpose for which it is used.ā
7.
ā¢TYPES OF VALIDITY
ļ¬1.Face Validity:
- it is the extent to which the
measurement method appears āon its
faceā to measure the construct of interest.
-is done by examining the physical
appearance of the instrument to make it
readable and understandable
8.
ļ¬EXAMPLE:
ļ¬People might havenegative reactions
to an intelligence test that did not
appear to them to be measuring their
intelligence.
9.
ļ¬ 2. ContentValidity:
-it is the extent to which the
measurement method covers the entire
range of relevant behaviors, thoughts,
and feelings that define the construct
being measured.
-is done through a careful and critical
examination of the objectives of
assessment to reflect the curricular
objectives.
10.
3. Criterion-based Validity:
-itis the extent to which peopleās
scores are correlated with other variables
or criteria that reflect the same construct.
-is established statistically such
that a set of scores revealed by the
measuring instrument is correlated with the
scores obtained in another external
predictor or measure
11.
ļ¬Example:
ļ¬ An IQtest should correlate positively with
school performance.
ļ¬An occupational aptitude test should
correlate positively with work performance
12.
ā¢TYPES OF CRITERIONVALIDITY:
ļ¬ 3.1. Predictive Validity:
-describes the future performance of
an individual by correlating the sets of
scores obtained from two measures given
at a longer time interval
-when the criterion is something that
will happen or be assessed in the future,
this is called predictive validity.
13.
ļ¬ 3.2. ConcurrentValidity:
-describes the present status of the
individual by correlating the sets of
scores obtained from two measures
given at a close interval
-when the criterion is something that is
happening or being assessed at the
same time as the construct of interest, it
is called concurrent validity.
14.
ļ¬4. Construct Validity
-isestablished statistically by comparing
psychological traits or factors that
theoretically influence scores in a test
15.
TYPES OF CONSTRUCTVALIDITY:
ļ¬4.1 Convergent Validity
-is established if the instrument defines
another similar trait other than what it is
intended to measure.
E.g. Critical Thinking Test may be
correlated with Creative Thinking Test.
16.
ļ¬4.2 Divergent Validity
-is established if an instrument can
describe only the intended trait and not
the other traits.
E.g. Critical Thinking Test may not be
correlated with Reading Comprehension
Test.
17.
Nature of Validity
1.Validity refers to the appropriateness of the
test results but not to the instrument itself.
2. Validity does not exist on an all-or-none
basis but it is a matter of degree.
3. Tests are not valid for all purposes. Validity
is always specific to particular interpretation.
4. Validity is not of different types. It is a unitary
concept. It is based on various types of
evidence.
18.
Factors Affecting Validity:-
1. Factors in the test:
(i) Unclear directions to the students to
respond the test.
(ii) Difficulty of the reading vocabulary and
sentence structure.
(iii) Too easy or too difficult test items.
(iv) Ambiguous statements in the test
items.
(v) Inappropriate test items for measuring
a particular outcome.
(vi) Inadequate time provided to take the
test
19.
2. Factors inTest Administration and
Scoring
(i) Unfair aid to individual students, who
ask for help.
(ii) Cheating by the pupils during testing.
(iii) Unreliable scoring of essay type
answer.
(iv) Insufficient time to complete the test.
(v) Adverse physical and psychological
condition at the time of testing.
20.
3. Factors relatedto Testee
(i) Test anxiety of the students.
(ii) Physical and Psychological state of
the pupil
(iii) Response setā a consistent
tendency to follow a certain pattern in
responding the items.
⢠RELIABILITY
ļ” refersto the consistency of
measurement; that is, how consistent
test results or other assessment results
from one measurement to another
23.
Other definitions givenby experts
ļ¬Gronlund and Linn (1995)-ā reliability
refers to the consistency of measurement-
that is, how consistent test scores or other
evaluation results are from one
measurement to otherā.
ļ¬Ebel and Frisbie (1991)- ā the term
reliability means the consistency with
which a set of test scores measure
whatever they do measureā.
24.
ļ¬C.V. Good (1973)-hasdefined reliability
as the ā worthiness with which a
measuring device measures something;
the degree to which a test or other
instrument of evaluation measures
consistently whatever it does in fact
measureā.
ļ¬Davis (1946) ā the degree of relative
precisions of measurement of a set of test
score is defined reliabilityā.
25.
Nature of Reliability
1.Reliability refers to consistency of the results
obtained with an instrument but not the
instrument itself.
2. Reliability refers to a particular interpretation
of test scores.
3.Reliability is a statistical concept to determine
reliability we administer a test to a group once
or more than once.
4. Reliability is necessary but not a sufficient
condition for validity.
26.
Four methods ofdetermining
reliability
ļ¬(a) Test-Retest method.
ļ¬(b) Equivalent forms/Parallel forms
method.
ļ¬(c) Split-half method.
ļ¬(d) Rational Equivalence/Kuder-
Richardson method.
27.
Test-Retest method:
ļ¬This isthe simplest method of determining
the test reliability.
ļ¬To determine reliability in this method the
test is given and repeated on same group.
Then the correlation between the first set
of scores and second set of scores is
obtained.
28.
Equivalent Forms/Parallel FormsMethod:
ļ¬In this process two parallel forms of tests
are administered to the same group of
pupils in short interval of time, then the
scores of both the tests are corĀrelated.
This correlation provides the index of
equivalence.
29.
Split-Half Method:
ļ¬In thismethod a test is administered to a
group of pupils in usual manner. Then the
test is divided into two equivalent values
and correlation for these halfĀtests are
found.
30.
Rational Equivalent/Kuder RichardsonMethod:
ļ¬This method also provides a measure of
internal consistency. It neither requires
administration of two equivalent forms of
tests nor it requires to split the tests into two
equal halves.
ļ¬Reliability coefficient is determined by using
the KuderĀRichardson formulaĀ20 which
reads like this.
31.
Factors affecting reliability:-
1.Factors related to test:
(i) length of the test
(ii) content of the test
(iii) characteristics of items
(iv) spread of scores
32.
2. Factors relatedto testee:
(i) Heterogeneity of the group
(ii) Test wiseness of the students
(iii) Motivation of the students
3. Factors related to testing procedures:
(i) Time limit of test
(ii) Cheating opportunity given to the
students
⢠OBJECTIVITY
ļ” refersto the agreement of two or more
raters or test administrators concerning
the score of the student.
35.
Other definitions givenby experts:
ļ¬C.V. Good (1973) defines objectivity in
testing is āthe extent to which the
instrument is free from personal error
(personal bias), that is subjectivity on the
part of the scorerā.
36.
ļ¬Gronlund and Linn(1995) states
āObjectivity of a test refers to the degree to
which equally competent scores obtain the
same results. So a test is considered
objective when it makes for the elimination
of the scorerās personal opinion and bias
judgement. In this conĀtext there are two
aspects of objectivity which should be kept
in mind while constructing a test.ā
37.
Two aspects ofobjectivity which should be kept
in mind while constructing a test
ļ¬1. Objectivity in Scoring
-means same person or different persons
scoring the test at any time arrives at the
same result without may chance error. Āa
test to be objective must necessarily so
worded that only correct answer can be
given to it.
38.
ļ¬2. Objectivity ofTest Items
-means that the item must call for a
definite single answer. WellĀconĀstructed
test items should lead themselves to one
and only one interpretation by students
who know the material involved. It means
the test items should be free from
ambiguity.
⢠FAIRNESS
- meansthe test item should not have
any biases. It should not be offensive
to any examinee subgroup.
-a test can only be good if it is fair to all
the examinees
-a fair assessment provides all
students with an equal opportunity to
demonstrate achievement
41.
The key tofairness are as follows:
ļ¬Students have knowledge of learning targets
and assessment.
ļ¬Students are given equal opportunity to
learn.
ļ¬Students possess the pre-requisite
knowledge and skills.
ļ¬Students are free from teacher stereotypes.
ļ¬Students are free from biased assessment
task and procedures.
⢠SCORABILITY
- meansthat the test should be easy
to score, direction for scoring should be
clearly stated in the instruction. Provide the
students an answer sheet and the answer
key for the one who will check the test.
⢠ADEQUACY
-means thatthe test should contain a
wide range of sampling of items to
determine the educational outcomes or
abilities so that the resulting scores are
representatives of the total performance in
the areas measured.
⢠ADMINISTRABILITY
- meansthat the test should be
administered uniformly to all students so
that the scores obtained will not vary due to
factors other than differences of the
studentās knowledge ad skills. There should
be a clear provision for instruction for the
students, proctors and even the one who will
check the test or the test scorer.
⢠PRACTICALITY ANDEFFICIENCY
-refers to the teacherās familiarity with
the methods used, time required for the
assessment, complexity of the
administration, ease of scoring, ease of
interpretation of the test results and the
materials used must be at the lowest cost.
⢠BALANCE
-a balancedassessment sets targets in all
domains of learning (cognitive, affective and
psychomotor) or domains of intelligence
(verbal-linguistic, logical-mathematical, bodily
kinesthetic , visual- spatial, musical-rhythmic,
intrapersonal-social, interpersonal-
introspection, physical world-natural,
existential-spiritual)
-makes use of both traditional and
alternative assessment