Professional Documents
Culture Documents
Psy Ass M1 Reviewer
Psy Ass M1 Reviewer
psychological assessment- gathering and integration of psychology - related data for the purpose of
making a psychological evaluation that is accomplished through the use of tools such as tests, interviews,
case studies, behavioral observation, and specially designed apparatuses and measurement procedures.
Scale
Battery
- term often used in test titles. A battery is a group of several tests, or subtests, that are
administered at one time to one person.
Process of assessment:
TEST
INTERVIEW
- Keep files of their work products (paper, canvas, film, video, audio)
- Sample of one’s ability and accomplishment
- Tool of evaluation
- Records, transcripts, and other accounts in written, pictorial or other form that preserve archival
info, official & informal accounts relevant to an assessee.
BEHAVIORAL OBSERVATION
- Monitoring the action of others or oneself by visual electronic means while recording
quantitative /qualitative information regarding the actions
- Venture outside of the confines of clinics, classrooms, workplaces, and research laboratories in
order to observe behaviour of humans in natural setting
ROLE-PLAY TESTS
COMPUTER AS TOOLS
A. Educational settings
B. Clinical settings
C. Counselling settings
D. Geriatric settings: Wherever older individuals reside, they may at some point require
psychological assessment to evaluate cognitive, psychological, adaptive, or other functioning
E. Business and military settings
F. Governmental & organizational credentialing
1. Responsible user have obligations before, during and after test or any measurement procedure is
administered
2. Test user must ensure that the room will be conducted, suitable and conducive to the settings
3. Attempts to establish rapport with the test taker not compromise any rules of the test
admiration instructions
4. Safeguard the test protocols & conveying the results in a clearly understandable fashion
Alternate assessment- diagnostic procedure varies from the usual, customary, or standardized way a
measurement is dervived either by virtue of some special accommodation made to the assessee or
by means of alternative methods designed to measure the same variables
Test catalogues
- Contain only brief description of the test and seldom contain the kind of detailed technical
information that a prospective user might require. The objective is to sell the test.
Test manuals
- Detailed information concerning the development of a particular test and technical information
relating to it should be found in the test manual
Reference volumes
Journal articles
- May contain reviews of the test, updated or independent studies of its psychometric soundness,
examples of how the instrument was used
- Rich source of information on important trends in testing and assessment
Online databases
HISTORICAL PERSPECTIVE:
Binet and Simon developed the first major general intelligence test.
Representative sample: one that comprises individuals similar to those for whom the test is to be used.
Mental age: a measurement of a child’s performance on the test relative to other children of that
particular age group
MODULE 2
Culture- socially transmitted behavior patterns, beliefs, and products of work of a particular population,
community, or group of people
LEVEL A: test that can be administered, scored, and interpreted with the aid of the manual and a
general orientation to the kind of institution or organization
LEVEL B: Require some technical knowledge of test construction and use and of supporting psychological
and educational fields such as statistics, individual differences, psychology of adjustment, personnel
psychology, and guidance.
LEVEL C: test that require substantial understanding of testing and supporting psychological fields
together with supervised experience in the use of these services
1. The right of informed consent - have a right to know why they are being evaluated, how the test
data will be used, and what (if any) information will be released to whom. With full knowledge
of such information, testtakers give their informed consent to be tested. disclosure of the
information needed for consent must, of course, be in language the testtaker can understand.
Consent must be in written rather than oral form. The written form should specify: general
purpose of the testing, specific reason it is being undertaken in the present case, general type of
instruments to be administered.
2. The right to be informed of test findings- Testtakers have a right to be informed, in language
they can understand, of the nature of the findings with respect to a test they have taken. In no
case would they disclose diagnostic conclusions that could arouse anxiety or precipitate a crisis.
giving realistic information about test performance to examinees is not only ethically and legally
mandated but may be useful from a therapeutic perspective as well. Testtakers have a right to be
informed, in language they can understand, of the nature of the findings with respect to a test
they have taken.
3. The right to privacy and confidentiality - Privilege in the psychologist-client relationship belongs
to the client, not the psychologist. The competent client can direct the psychologist to disclose
information to some third party (such as an attorney or an insurance carrier), and the
psychologist is obligated to make the disclosure. In some rare instances, the psychologist may be
ethically (if not legally) compelled to disclose information if that information will prevent harm
either to the client or to some endangered third party. Clinicians may have a duty to warn
endangered third parties not only of potential violence but of potential AIDS infection from an
HIV-positive client as well as other threats to their physical well-being.
4. The right to the least stigmatizing label - least stigmatizing labels should always be assigned
when reporting test results.
For assessment professionals, some major issues with regard to Computer-assisted
psychological assessment (CAPA) are as follows:
MODULE 3
Measurement- phenomenon that is measured is made more easily subject to confirmation and analysis,
and thus is made more objective as well
discrete variables are those with a finite range of values—or a potentially infinite, but countable, range
of values. (e.g. Dichotomous, polytomous)
Continuous variables such as time, distance, and temperature, on the other hand, have infinite ranges
and really cannot be counted; can never be calibrated with enough precision
SCALES OF MEASUREMENT
Nominal: classification or categorization; numbers are used solely as labels; Numbers can also be used to
label categorical data; results of such operations are not meaningful.
Ordinal: ordinal scales permit classification; rank ordering; e elements in a set can be lined up in a series
—from lowest to highest; they carry no information with regard to the distance between positions.
Interval: contain equal intervals between numbers; Each unit on the scale is exactly equal to any other
unit on the scale.
Ratio: has a true zero point. All mathematical operations can meaningfully be performed because there
exist equal intervals; In psychology, ratio-level measurement is employed in some types of tests and test
items, perhaps most notably those involving assessment of neurological functioning; , all with
meaningful results.
TYPES OF STATISTICS
Descriptive: Numbers and graphs used to describe, condense, or represent data belong in the realm of
descriptive statistics
Inferential: when data are used to estimate population values based on sample values or to test
hypotheses
- Frequency distribution: organize raw data in some sensible way so they can be inspected; help
to organize scores into a still more compact form; , scores are grouped into intervals of a
convenient size to accommodate the data, and the frequencies are listed for each interval
instead of for each of the scores; simple frequency distribution to indicate that individual scores
have been used and the data have not been grouped
- Graphs: a diagram or chart composed of lines, points, bars, or other symbols that describe and
illustrate data.
- Measures of central tendency- statistic that indicates the average or midmost score between
the extreme scores in a distribution.
- Arithmetic mean- e most appropriate measure of central tendency for interval or ratio data
when the distributions are believed to be approximately normal
MEASURES OF VARIABILITY
range: range of a distribution is equal to the difference between the highest and the lowest scores; e
simplest measure of variability to calculate; range provides a quick but gross description of the spread of
scores. When its value is based on extreme scores in a distribution, the resulting description of variation
may be understated or overstated
In a perfectly symmetrical distribution, Q 1 and Q 3 will be exactly the same distance from the median
If these distances are unequal then there is a lack of symmetry. This lack of symmetry is referred to as
skewness
All the deviation scores are then summed and divided by the total number of scores ( n ) to arrive at the
average deviation.
Standard deviation - as a measure of variability equal to the square root of the average squared
deviations about the mean. More succinctly, it is equal to the square root of the variance
- must be calculated same as the average deviation but Instead of using the absolute value of each
deviation score, we use the square of each score.
variance is equal to the arithmetic mean of the squares of the differences between the scores in a
distribution and their mean
positive skew when relatively few of the scores fall at the high end of the distribution. Positively skewed
examination results may indicate that the test was too difficult
standard score is a raw score that has been converted from one scale to another scale, where the latter
scale has some arbitrarily set mean and standard deviation
z score is equal to the difference between a particular raw score and the mean divided by the standard
deviation
MODULE 4
- trait has been defined as “any distinguishable, relatively enduring way in which one individual
varies from another”.
- States also distinguish one person from another but are relatively less enduring.
- construct — an informed, scientific concept developed or constructed to describe or explain
behavior.
- overt behavior refers to an observable action or the product of an observable action, including
test- or assessment-related responses.
- test score is presumed to represent the strength of the targeted ability or trait or state and is
frequently based on cumulative scoring
- the more the testtaker responds in a particular direction as keyed by the test manual as correct
or consistent with a particular trait, the higher that testtaker is presumed to be on the targeted
ability or trait.
- Patterns of answers to true–false questions on one widely used test of personality are used in
decision making regarding mental disorders.
- tasks in some tests mimic the actual behaviors that the test user is attempting to understand.
such tests yield only a sample of the behavior that can be expected to be emitted under nontest
conditions.
- obtained sample of behavior is typically used to make predictions about future behavior, such as
work performance of a job applicant.
Assumption 4: Tests and Other Measurement Techniques Have Strengths and Weaknesses
- Competent test users understand and appreciate the limitations of the tests they use as well as
how those limitations might be compensated for by data from other sources
- error refers to a long-standing assumption that factors other than what a test attempts to
measure will influence performance on the test
- error variance- component of a test score attributable to sources other than the trait or ability
measured.
Assumption 6: Testing and Assessment Can Be Conducted in a Fair and Unbiased Manner
- all major test publishers strive to develop instruments that are fair when used in strict
accordance with guidelines in the test manual.
- In a world without tests or other assessment procedures, personnel might be hired on the basis
of nepotism rather than documented merit
psychometric soundness- technical criteria that assessment professionals use to evaluate the quality of
tests and other measurement procedures.
Reliability- consistency of the measuring tool: the precision with which the test measures and the extent
to which error is present in measurements.
MODULE 5
Test construction: item sampling or content sampling, terms that refer to variation among items within a
test as well as to variation among items between tests
Test administration: may influence the testtaker’s attention or motivation. The testtaker’s reactions to
those influences are the source of one kind of error variance. Other potential sources of error variance
during test administration are testtaker variables. . Examiner-related variables are potential sources of
error variance.
RELIABILITY ESTIMATES:
1. Test-retest reliability
2. Parallel forms & alternate forms
3. Split-half
4. Inter-scorer
Test scoring & interpretation: The advent of computer scoring and a growing reliance on objective,
computer-scorable items virtually have eliminated error variance caused by scorer differences in many
tests.
o The Spearman-Brown formula allows a test developer or user to estimate internal consistency
reliability from a correlation of two halves of a test.
o Inter-item consistency refers to the degree of correlation among all the items on a scale.
- Tests are said to be homogeneous if they contain items that measure a single trait
- KR-20 is the statistic of choice for determining the inter-item consistency of
dichotomous items, primarily those items that can be scored right or wrong (such as
multiple-choice items)
- Coefficient alpha is the preferred statistic for obtaining an estimate of internal
consistency reliability. Coefficient alpha is widely used as a measure of reliability, in part
because it requires only one administration of the test.
o Inter-scorer reliability is thedegree of agreement or consistency between two or more scorers
(or judges or raters) with regard to a particular measure.
MODULE 6 VALIDITY
Validity - judgment based on evidence about the appropriateness of inferences drawn from test scores.
Face validity
Content validity
Criterion-related validity
- how adequately a test score can be used to infer an individual’s most probable standing on some
measure of interest
Concurrent validity
- index of the degree to which a test score is related to some criterion measure obtained at the
same time
- extent to which test scores may be used to estimate an individual’s present standing on a
criterion.
Predictive validity
Incremental validity
- Test users involved in predicting some criterion from test scores are often interested in the utility
of multiple predictors. The value of including more than one predictor depends on a couple of
factors
- Expectancy data provide information that can be used in evaluating the criterion-related validity
of a test.
Construct validity
- t the appropriateness of inferences drawn from test scores regarding individual standings on a
variable called a construct
Test bias
- having to do with prejudice and preferential treatment.
Rating error
- numerical or verbal judgment (or both) that places a person or an attribute along a continuum
identified by a scale of numerical or word descriptors known as a rating scale
- intentional or unintentional misuse of rating scale
A leniency error (also known as a generosity error ) is, as its name implies, an error in rating that
arises from the tendency on the part of the rater to be lenient in scoring, marking, and/or
grading
- central tendency error. Here the rater, for whatever reason, exhibits a general and systematic
reluctance to giving ratings at either the positive or the negative extreme. Consequently, all of
this rater’s ratings would tend to cluster in the middle of the rating continuum
- Halo effect rater’s failure to discriminate among conceptually distinct and potentially
independent aspects of a ratee’s behavior
MODULE 7 UTILITY