Psychometrics # MCQs Practice set

Q.1 What does the term 'reliability' in psychometrics refer to?

Accuracy of measurement
Consistency of measurement
Validity of the construct
Difficulty of test items
Explanation - Reliability refers to the consistency or stability of test scores over repeated applications.
Correct answer is: Consistency of measurement

Q.2 Which of the following is an example of a norm-referenced test?

IQ test
Driving license test
Certification exam
Personality questionnaire
Explanation - Norm-referenced tests compare an individual's performance to a group (norm) to interpret scores.
Correct answer is: IQ test

Q.3 Cronbach's alpha is used to measure:

Inter-rater reliability
Internal consistency
Test-retest reliability
Construct validity
Explanation - Cronbach's alpha measures how closely related a set of items are as a group, reflecting internal consistency.
Correct answer is: Internal consistency

Q.4 A test is said to have high construct validity if it:

Measures what it claims to measure
Has consistent scores
Has few questions
Is easy to administer
Explanation - Construct validity evaluates whether a test truly measures the theoretical construct it is intended to measure.
Correct answer is: Measures what it claims to measure

Q.5 Which type of reliability is assessed by giving the same test to the same group twice?

Parallel forms reliability
Test-retest reliability
Split-half reliability
Inter-rater reliability
Explanation - Test-retest reliability measures the stability of test scores over time by administering the same test twice.
Correct answer is: Test-retest reliability

Q.6 In item analysis, an item with very low discrimination index:

Differentiates well between high and low scorers
Fails to differentiate between high and low scorers
Is highly correlated with the test total
Should always be retained
Explanation - Items with low discrimination index do not distinguish between those who have mastered the content and those who haven't.
Correct answer is: Fails to differentiate between high and low scorers

Q.7 The primary purpose of standardization in testing is to:

Make a test easier
Ensure consistent administration and scoring
Reduce test length
Increase creativity in responses
Explanation - Standardization ensures all test-takers experience the test under the same conditions, making scores comparable.
Correct answer is: Ensure consistent administration and scoring

Q.8 Which scale of measurement allows for ranking but not precise differences between scores?

Nominal
Ordinal
Interval
Ratio
Explanation - Ordinal scales rank order items but do not indicate the magnitude of difference between them.
Correct answer is: Ordinal

Q.9 A test measuring multiple abilities, such as verbal and numerical skills, is called a:

Unidimensional test
Multidimensional test
Criterion-referenced test
Speed test
Explanation - Multidimensional tests assess more than one construct or ability, unlike unidimensional tests which focus on a single trait.
Correct answer is: Multidimensional test

Q.10 Face validity refers to:

Actual statistical correlation with other measures
Appearance of measuring what it intends to measure
Consistency over time
Difficulty level of items
Explanation - Face validity is subjective and refers to whether a test looks like it measures the intended construct, based on judgment.
Correct answer is: Appearance of measuring what it intends to measure

Q.11 Which method is commonly used to estimate inter-rater reliability?

Pearson correlation
Cohen’s kappa
Cronbach’s alpha
Factor analysis
Explanation - Cohen’s kappa measures agreement between two raters beyond chance, making it suitable for categorical ratings.
Correct answer is: Cohen’s kappa

Q.12 What is a percentile rank in psychometrics?

The raw score of a test
The percentage of people scoring below a given score
The number of correct responses
The mean score of a sample
Explanation - Percentile rank indicates the relative standing of a score within a group by showing the percentage of test-takers scoring lower.
Correct answer is: The percentage of people scoring below a given score

Q.13 A criterion-referenced test differs from a norm-referenced test because it:

Compares scores to a standard
Compares scores to a group
Always has multiple-choice items
Cannot measure ability accurately
Explanation - Criterion-referenced tests measure performance against a fixed standard or learning objective, not against other test-takers.
Correct answer is: Compares scores to a standard

Q.14 Item difficulty in a test is usually represented by:

Proportion of test-takers answering correctly
Standard deviation of scores
Cronbach’s alpha
Factor loadings
Explanation - Item difficulty index (p-value) indicates how many test-takers answered the item correctly; higher values mean easier items.
Correct answer is: Proportion of test-takers answering correctly

Q.15 Which of the following best defines construct underrepresentation?

Test fails to include important aspects of the construct
Test measures irrelevant content
Test is too easy
Test has low reliability
Explanation - Construct underrepresentation occurs when a test does not fully capture the intended construct, limiting its validity.
Correct answer is: Test fails to include important aspects of the construct

Q.16 Split-half reliability involves:

Dividing the test into two halves and correlating scores
Administering two versions of the test
Relying on a single rater
Comparing scores to a standard
Explanation - Split-half reliability estimates internal consistency by splitting a test and correlating the two halves.
Correct answer is: Dividing the test into two halves and correlating scores

Q.17 A latent variable is:

Directly observable
Inferred from measured variables
Always numerical
The same as a raw score
Explanation - Latent variables are constructs like intelligence or anxiety, which are not directly measurable but inferred through tests.
Correct answer is: Inferred from measured variables

Q.18 Test-retest reliability can be affected by:

Memory effects, learning, and time interval
Item difficulty only
Number of items only
Sample size only
Explanation - Scores may change due to memory of items or actual learning between test administrations, impacting reliability.
Correct answer is: Memory effects, learning, and time interval

Q.19 Which type of validity is assessed by comparing a test to another established test measuring the same construct?

Content validity
Concurrent validity
Predictive validity
Face validity
Explanation - Concurrent validity assesses whether a test correlates with an already established measure of the same construct.
Correct answer is: Concurrent validity

Q.20 A test with low reliability will generally have:

High validity
Unstable scores
High discrimination
Large sample size
Explanation - Low reliability indicates inconsistency in scores, making results less trustworthy.
Correct answer is: Unstable scores

Q.21 In factor analysis, 'factor loadings' represent:

Correlation between observed variables and latent factors
Number of items in a test
Test difficulty
Test reliability
Explanation - Factor loadings indicate how strongly each observed variable relates to an underlying factor in factor analysis.
Correct answer is: Correlation between observed variables and latent factors

Q.22 Which type of test is designed to predict future performance?

Achievement test
Aptitude test
Norm-referenced test
Personality test
Explanation - Aptitude tests aim to measure potential or ability to succeed in future tasks, unlike achievement tests which measure current knowledge.
Correct answer is: Aptitude test

Q.23 Content validity is primarily established through:

Expert judgment about the test items
Statistical correlation with other tests
Test-retest reliability
Cronbach's alpha
Explanation - Content validity is determined by experts who evaluate whether test items adequately cover the construct domain.
Correct answer is: Expert judgment about the test items

Q.24 A ceiling effect occurs when:

Most test-takers achieve the highest possible score
Most test-takers score very low
Test scores are normally distributed
Test items are too diverse
Explanation - Ceiling effects occur when a test is too easy, limiting differentiation among high scorers.
Correct answer is: Most test-takers achieve the highest possible score

Q.25 Which is the best description of criterion-related validity?

Extent to which a test correlates with an outcome measure
Degree to which test items appear relevant
Consistency of test scores
Internal structure of the test
Explanation - Criterion-related validity evaluates how well a test predicts outcomes based on a criterion measure, either concurrently or in the future.
Correct answer is: Extent to which a test correlates with an outcome measure