The effects of response instructions on situational judgment test performance and validity in a high-stakes context
This study fills a key gap in research on response instructions in situational judgment tests (SJTs). The authors examined whether the assumptions behind the differential effects of knowledge and behavioral tendency SJT response instructions hold in a large-scale high-stakes selection context (i.e.,...
Saved in:
Main Authors: | , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2009
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/lkcsb_research/5670 https://ink.library.smu.edu.sg/context/lkcsb_research/article/6669/viewcontent/SJTinstructions.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.lkcsb_research-6669 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.lkcsb_research-66692019-08-23T03:05:05Z The effects of response instructions on situational judgment test performance and validity in a high-stakes context LIEVENS, Filip SACKETT, Paul R. BUYSE, Tine This study fills a key gap in research on response instructions in situational judgment tests (SJTs). The authors examined whether the assumptions behind the differential effects of knowledge and behavioral tendency SJT response instructions hold in a large-scale high-stakes selection context (i.e., admission to medical college). Candidates (N = 2,184) were randomly assigned to a knowledge or behavioral tendency response instruction SJT, while SJT content was kept constant. Contrary to prior research in low-stakes settings, no meaningfully important differences were found between mean scores for the response instruction sets. Consistent with prior research, the SJT with knowledge instructions correlated more highly with cognitive ability than did the SJT with behavioral tendency instructions. Finally, no difference was found between the criterion-related validity of the SJTs under the two response instruction sets. 2009-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/lkcsb_research/5670 info:doi/10.1037/a0014628 https://ink.library.smu.edu.sg/context/lkcsb_research/article/6669/viewcontent/SJTinstructions.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection Lee Kong Chian School Of Business eng Institutional Knowledge at Singapore Management University Situational judgment test response instructions high-stakes testing Human Resources Management Organizational Behavior and Theory |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Situational judgment test response instructions high-stakes testing Human Resources Management Organizational Behavior and Theory |
spellingShingle |
Situational judgment test response instructions high-stakes testing Human Resources Management Organizational Behavior and Theory LIEVENS, Filip SACKETT, Paul R. BUYSE, Tine The effects of response instructions on situational judgment test performance and validity in a high-stakes context |
description |
This study fills a key gap in research on response instructions in situational judgment tests (SJTs). The authors examined whether the assumptions behind the differential effects of knowledge and behavioral tendency SJT response instructions hold in a large-scale high-stakes selection context (i.e., admission to medical college). Candidates (N = 2,184) were randomly assigned to a knowledge or behavioral tendency response instruction SJT, while SJT content was kept constant. Contrary to prior research in low-stakes settings, no meaningfully important differences were found between mean scores for the response instruction sets. Consistent with prior research, the SJT with knowledge instructions correlated more highly with cognitive ability than did the SJT with behavioral tendency instructions. Finally, no difference was found between the criterion-related validity of the SJTs under the two response instruction sets. |
format |
text |
author |
LIEVENS, Filip SACKETT, Paul R. BUYSE, Tine |
author_facet |
LIEVENS, Filip SACKETT, Paul R. BUYSE, Tine |
author_sort |
LIEVENS, Filip |
title |
The effects of response instructions on situational judgment test performance and validity in a high-stakes context |
title_short |
The effects of response instructions on situational judgment test performance and validity in a high-stakes context |
title_full |
The effects of response instructions on situational judgment test performance and validity in a high-stakes context |
title_fullStr |
The effects of response instructions on situational judgment test performance and validity in a high-stakes context |
title_full_unstemmed |
The effects of response instructions on situational judgment test performance and validity in a high-stakes context |
title_sort |
effects of response instructions on situational judgment test performance and validity in a high-stakes context |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2009 |
url |
https://ink.library.smu.edu.sg/lkcsb_research/5670 https://ink.library.smu.edu.sg/context/lkcsb_research/article/6669/viewcontent/SJTinstructions.pdf |
_version_ |
1770574051296673792 |