100 Questions
Which type of test primarily relies on content validity?
Achievement test
What is the primary focus of an aptitude test?
Potential for acquiring specific skills
Which type of personality test involves ambiguous stimuli or responses?
Projective personality test
What is an example of a typical performance test?
Personality test
Which of the following is NOT a type of structured personality test?
Ambiguous stimulus test
Which property is essential in a psychological test and pertains to its technical quality?
Psychometric Soundness
What type of interview method allows subjects to express their feelings freely without fear of disapproval?
Non-Directive
Which type of neuropsychological test is designed to measure intervals between events and behaviors?
Typical performance test
What type of psychological assessment involves the assessor and assessee working as 'partners'?
Collaborative
What differentiates psychological testing from psychological assessment?
Testing requires technician-like skills, while assessment requires expert selection and integration of tools.
Which component of non-directive interviews involves biographical sketch of the client?
Social case
What is the primary purpose of a psychological test item?
To elicit a specific response for scoring or evaluation
Which type of test measures one's likes, dislikes, and personality orientation towards the world of work?
Interest inventory
What is the main objective of a motivational interview?
Gathering information and addressing problematic behaviors therapeutically
Which type of assessment is conducted without physical proximity between the subject and evaluator?
Remote
What is a 'cut-score' in the context of psychological testing?
A reference point used to classify data
Which assessment approach encourages therapeutic self-discovery?
Therapeutic
Which component of a psychological test refers to its form, plan, and structure?
Format
In psychological assessment, which approach involves evaluating specific problems at the moment they occur?
Ecological Momentary
What primarily distinguishes a psychological trait from a psychological state?
Traits are enduring while states are transient
Which level of interpretation involves minimal concern with intervening processes and no concern with underlying constructs?
Level I
In psychological testing, what does 'scoring' entail?
Assigning scores to test performances
In the context of psychological assessment, what is Extra-Test Behavior?
Observations regarding examinee's reactions during testing
Which of the following best describes Mechanical Prediction?
Application of computer algorithms and statistical rules
Who is primarily responsible for creating the tests or other methods of assessment?
Test Authors/Developers
In actuarial assessment, what predominately influences the assessor's judgment?
Statistical rules
Which level of interpretation involves the development of a coherent and inclusive theory of the individual's life?
Level III
What is one of the main purposes of psychological traits, according to the assumptions?
To predict the present from the past
Which of the following is true about hypothetical constructs in psychological testing?
They explain visible behavior based on assumed inner states
In psychological assessment, who controls the distribution of tests?
Test Publishers
What type of variance is attributable to irrelevant random sources?
Error Variance
Which type of error can consistently affect the true score without changing the standard deviation?
Systematic Error
What term refers to variation among items within a test as well as to variation among items between tests?
Item Sampling/Content Sampling
What is one effect of having a short test-retest interval?
Carryover Effects
What is meant by 'practice effect' in the context of test-retest reliability?
Scores are higher in the second session due to experience from the first session
What is a common result when the time span between test-retest occasions is lengthy?
Insignificant reliability coefficient
When does test sophistication primarily occur?
When test items are challenging and remembered by the test takers
Which of the following contributes to inconsistency in test scores?
Characteristics of the test situation
What factor is NOT typically associated with measurement error?
Stable attributes of the individual
Which formula is used to calculate the true score?
$True Score = R_{xx}(ar{x} - ar{x}) + ar{x}$
Which of the following accurately describes the relationship between reliability and the Standard Error of Measurement (SEM)?
Higher reliability, lower SEM
Which of the following is NOT used to increase the internal consistency of a test?
Implementing face validity
What does the Standard Error of the Difference help determine?
How large a difference should be before it is considered statistically significant
Which term refers to the generalizability of research results?
External validity
What does a higher confidence level in a confidence interval imply?
Larger range of possible true scores
What aspect of validity is concerned with the degree to which test items represent a target domain?
Content validity
What does test sensitivity specifically detect?
True positives
What outcome is characterized by predicting success that does not occur?
False Positive
What does the term face validity refer to?
A test's ability to appear to measure what it actually measures to the person being tested
Which statement best describes internal validity?
The degree of control among variables in a study
Which method assesses agreement between two or more raters with a categorical assessment scale?
Fleiss Kappa
What does the Domain Sampling Theory aim to estimate?
The extent to which specific sources of variation contribute to test scores
In Item Response Theory, what does the term 'Discrimination' refer to?
The degree to which an item differentiates among people with varying levels of a trait
Which theory is concerned with the variables in testing situations causing score variations?
Generalizability Theory
Which type of test has a time limit long enough to allow test takers to attempt all questions?
Power Test
Standard Error of Measurement is designed to provide what?
A precision measure of an observed test score
What does the 'Decision Study' in Generalizability Theory examine?
The usefulness of test scores in decision-making
What does a high degree of internal consistency in a homogenous test indicate?
The test is designed to measure a single factor
Which type of error provides an index of the expected error in an individual's test score?
Standard Error of Measurement
According to Generalizability Theory, what would result in the exact same test score being obtained?
Given the exact same conditions of all the facets in the testing universe
Who developed the methodology of factor analysis?
Charles Spearman
Which type of factor analysis involves testing the degree to which a hypothetical model fits the actual data?
Confirmatory Factor Analysis
Which type of rating error involves the tendency to cluster ratings in the middle of the scale?
Central Tendency Error
What term is used to describe the revalidation of a test to a criterion based on a different group?
Cross-Validation
Which method is used to overcome rating errors by using rankings instead of ratings?
Ranking
What is the name of the table that provides an estimate of the usefulness of a test in a selection system?
Taylor-Russell Table
Which type of validity is assessed by using factor analysis to study the interrelationships among a set of variables?
Convergent and Discriminant Validity
What is the term for preferential treatment that prevents impartial measurement in a test?
Prejudice
Which error occurs when a rater is excessively strict in scoring?
Severity Error
What basic element in utility analysis is described as both economic and noneconomic terms?
Cost
What is the primary concern of a test blueprint in psychological assessment?
Its representativeness of a defined body of content
Which type of validity is concerned with the use of a test score to infer an individual's most probable standing on some measure of interest?
Criterion Validity
What does a zero Content Validity Ratio (CVR) indicate according to Lawshe's formula?
Exactly half of the experts rate the item as essential
What is the main objective of examining incremental validity in the context of psychological testing?
To explain the criterion measure better than current predictors
Which of the following is a method for improving the homogeneity of a test containing dichotomous items?
Eliminate items that do not show significant correlation coefficients with total test scores
What is a primary characteristic of a criterion in validity testing?
Being a relevant, valid and uncontaminated standard
What does construct-irrelevant variance refer to?
Variability in test scores due to factors unrelated to the construct
Which type of evidence demonstrates that test scores vary predictably based on group membership?
Method of Contrasted Groups
Which type of validity evidence would show a low correlation between the test being validated and other unrelated variables?
Discriminant Evidence
What term describes a failure to capture important components of a construct in a test?
Construct Underrepresentation
What is the main purpose of cumulative scoring in psychological assessment?
To gauge the strength of the respondent's traits
According to Assumption 3, what does test-related behavior aim to predict?
Non-test-related behavior
What does the reliability coefficient measure?
Consistency of test scores over time
Which of the following is NOT considered a potential source of error variance?
Test subject's educational qualifications
In Classical Test Theory, what does 'E' represent in the formula $X = T + E$?
Error
Competent test users understand the limitations of tests and:
Compensate for them with data from other sources
How is psychological reliability defined?
Dependability or consistency of the instrument or scores
Which of the following best describes error variance?
Variance due to factors other than the trait being measured
What does it mean when a test is said to be free from errors?
It has high reliability
Error in psychological assessment refers to:
Influences that affects performance besides the targeted traits
What is the main challenge associated with Parallel Forms/Alternate Forms Reliability?
Difference between the two tests
Which error is specifically associated with Split-Half Reliability?
Item Sample: Nature of Split
Which of the following tools are used for evaluating the coefficient of stability?
Pearson R and Spearman Rho
Homogeneity in tests is crucial for which type of reliability measurement?
Internal Consistency (Inter-Item Reliability)
Under what condition should the KR-21 formula be used?
For tests with same degree of difficulty
Why is counterbalancing used in psychological assessments?
To avoid carryover effects in parallel forms
Which technique is suitable for testing internal consistencies of a test that contains non-dichotomous items?
Cronbach's Coefficient Alpha
When is it impractical or undesirable to assess reliability with two tests or to administer a test twice, which reliability is useful?
Split-Half Reliability
What statistical tool is used along with Pearson R for evaluating Parallel Forms/Alternate Forms Reliability?
Spearman Rho
Which measure focuses on the degree of differences that exist between item scores to evaluate internal consistencies?
Average Proportional Distance
Study Notes
Psychological Assessment
Definition of Psychological Testing
- process of measuring psychology-related variables by means of devices or procedures designed to obtain a sample of behavior
- numerical in nature, individual or by group, administrators can be interchangeable without affecting the evaluation, requires technician-like skills in terms of administration and scoring, yield a test score or series of test scores, minutes to few hours
Psychological Assessment
- gathering and integration of psychology-related data for the purpose of making psychological evaluation
- answers referral question thru the use of different tools of evaluation
- individual, assessor is the key to the process of selecting tests and/or other tools of evaluation, requires an educated selection of tools of evaluation, skill in evaluation, and thoughtful organization and integration of data, entails logical problem-solving that brings to bear many sources of data assigned to answer the referral question
Types of Psychological Assessments
Ecological Momentary Assessment
- "in the moment" evaluation of specific problems and related cognitive and behavioral variables at the very time and place that they occur
Collaborative Assessment
- the assessor and assessee may work as "partners" from initial contact through final feedback
Therapeutic Assessment
- therapeutic self-discovery and new understanding are encouraged
Dynamic Assessment
- describe interactive approach to psychological assessment that usually follows the model: evaluation > intervention of some sort > evaluation
Psychological Test
- device or procedure designed to measure variables related to psychology
- components: content, format, item, administration procedures, score, scoring, cut-score, psychometric soundness, psychometrics, psychometrist or psychometrician
Ability or Maximal Performance Test
- Assess what a person can do
- types: achievement, aptitude, intelligence
Typical Performance Test
- Measure usual or habitual thoughts, feelings, and behavior
- indicate how test takers think and act on a daily basis
- use interval scales, no right and wrong answers
Personality Test
- measures individual dispositions and preferences
- designed to identify characteristic
- types: structured, projective, attitude, interest inventories
Other Tests
- speed, power, values inventory, trade, neuropsychological, norm-referenced, criterion-referenced
Interview
- method of gathering information through direct communication involving reciprocal exchange
- types: standardized/structured, non-standardized/unstructured, semi-standardized/focused, non-directive
Non-Directive Interview Components
- mental status examination, intake interview, social case, employment interview, panel interview, motivational interview, portfolio, case history data, case study, groupthink, behavioral observation
Assumptions of Psychological Assessment
-
- Psychological traits and states exist
-
- Psychological traits and states can be quantified and measured
-
- Test-related behavior predicts non-test-related behavior
-
- Test and other measurement techniques have strengths and weaknesses
-
- Various sources of error are part of the assessment process
-
- Testing and assessment can be conducted in a fair and unbiased manner
-
- Testing and assessment benefit society
Reliability
- dependability or consistency of the instrument or scores obtained by the same person when re-examined with the same test on different occasions, or with different sets of equivalent items
- estimate the range of possible random fluctuations that can be expected in an individual's score
- free from errors, minimizing error, using only representative sample to obtain an observed score
- true score cannot be found, reliability coefficient, index of reliability
Classical Test Theory
- score on an ability test is presumed to reflect not only the testtaker's true score on the ability being measured but also the error
- error: refers to the component of the observed test score that does not have to do with the testtaker's ability
- errors of measurement are random
Data Collection and Interpretation
- data collection, data interpretation, hit rate, profile, actuarial assessment, mechanical prediction
Levels of Interpretation
- level I: minimal amount of any sort of interpretation
- level II: descriptive generalizations, hypothetical construct
- level III: the effort to develop a coherent and inclusive theory of the individual life or a "working image" of the patient
Parties in Psychological Assessment
- test author/developer, test publishers, test reviewers, test users, test takers, test sponsors, society### Reliability Estimates
- Provide an estimate of the amount of error inherent in an observed score or measurement
- Used to estimate or infer the extent to which an observed score deviates from a true score
- Higher reliability, lower Standard Error of Measurement (SEM)
Types of Reliability
Parallel Forms Reliability
- Established when at least two different versions of the test yield almost the same scores
- Has the most universal applicability
- Alternate Forms: simply different versions of a test constructed to be parallel
Internal Consistency (Inter-Item Reliability)
- Used when tests are administered once
- Measures the internal consistency of the test, which is the degree to which each item measures the same construct
- Error: Item Sampling Homogeneity
- Statistical Tools:
- KR-20 for inter-item consistency of dichotomous items
- KR-21 for tests with equal variances and dichotomous scores
- Cronbach's Coefficient Alpha for non-dichotomous items with unequal variances
- Average Proportional Distance for internal consistency of a test
Split-Half Reliability
- Obtained by correlating two pairs of scores obtained from equivalent halves of a single test administered once
- Useful when it is impractical or undesirable to assess reliability with two tests or to administer a test twice
Validity
- A judgment or estimate of how well a test measures what it is supposed to measure
- Inferences: logical result or deduction
- Degree to which the measurement procedure measures the variables to be measured
- Validation: the process of gathering and evaluating evidence about validity
- Types of Validity:
- Face Validity: a test appears to measure what it is supposed to measure
- Content Validity: a test represents a adequate sample of a defined universe of situations or behaviors
- Criterion Validity: a test score is used to infer an individual's most probable standing on a measure of interest
- Construct Validity (Umbrella Validity): a judgment about the appropriateness of inferences drawn from test scores regarding individual standing on a variable called construct
Error
Scorer Differences
- Degree of agreement or consistency between two or more scorers regarding a particular measure
- Evaluated by calculating the percentage of times that two individuals assign the same scores to the performance of the examinees
Systematic Error
- Error that results from a flaw or bias in the measurement instrument or testing procedure
Domain Sampling Theory
- A theory that estimates the extent to which specific sources of variation under defined conditions are contributing to the test scores
- Generalizability Theory: a theory that considers the problem created by using a limited number of items to represent a larger and more complicated construct
Item Response Theory
- A theory that focuses on the probability that a person with a certain ability will be able to perform at a certain level on a test
- Difficulty: an attribute of not being easily accomplished, solved, or comprehended
- Discrimination: the degree to which an item differentiates among people with higher or lower levels of the trait, ability, etc.
Standard Error of Measurement
- Provides a measure of the precision of an observed test score
- Index of the amount of inconsistent or the amount of the expected error in an individual's score
A review quiz for psychological assessment, covering topics such as psychometric properties and more. Based on sources from Cohen & Swerdlik, Kaplan & Saccuzzo, and Groth & Wright.
Make Your Own Quizzes and Flashcards
Convert your notes into interactive study material.
Get started for free