Assessment centers (ACs) are a popular evaluation approach often applied for the purposes of guiding employment selection and development decisions. AC participants are required to engage in a series of work simulation exercises (e.g., role plays, group discussions, and presentations), and their performance on those exercises is rated by trained assessors. It is these interactions between participants and work simulations that inspired early organizational interest in the AC approach in the late 1940s (Handyside & Duncan, Reference Handyside and Duncan1954; Highhouse & Nolan, Reference Highhouse, Nolan, Jackson, Lance and Hoffman2012) that has remained into the present.
ACs continue to hold appeal in contemporary organizations, which is likely due, in part, to the their interpersonal nature (Kleinmann & Ingold, Reference Kleinmann and Ingold2019) and the rich source of job-relevant information they provide, particularly on job candidates and for employee development (Lievens, Reference Lievens2009). In the same manner, ACs continue to motivate the interests of researchers, as evidenced by the volume of empirical articles on ACs published over the last 10 years (e.g., Breil et al., Reference Breil, Lievens, Forthmann and Back2023; Dimotakis et al., Reference Dimotakis, Mitchell and Maurer2017; Heimann et al., Reference Heimann, Ingold, Lievens, Melchers, Keen and Kleinmann2022; Hickman et al., Reference Hickman, Herde, Lievens and Tay2023; Hoffman et al., Reference Hoffman, Kennedy, LoPilato, Monahan and Lance2015; Ingold et al., Reference Ingold, Kleinmann, Konig and Melchers2016, Reference Ingold, Dönni and Lievens2018; Jackson et al., Reference Ingold, Kleinmann, Konig and Melchers2016, Reference Jackson, Michaelides, Dewberry, Nelson and Stephens2022; Jansen et al., Reference Jansen, Melchers, Lievens, Kleinmann, Brandli, Fraefel and Konig2013; Kuncel & Sackett, Reference Kuncel and Sackett2014; Lievens et al., Reference Lievens, Schollaert and Keen2015; Meriac et al., Reference Meriac, Hoffman and Woehr2014; Monahan et al., Reference Monahan, Hoffman, Lance, Jackson and Foster2013; Oliver et al., Reference Oliver, Hausdorf, Lievens and Conlon2016; Putka & Hoffman, Reference Putka and Hoffman2013; Sackett et al., Reference Sackett, Shewach and Keiser2017; Speer et al., Reference Speer, Christiansen, Goffin and Goff2014; Thornton et al., Reference Thornton, Rupp, Gibbons and Vanhove2019; Wirz et al., Reference Wirz, Melchers, Kleinmann, Lievens, Annen, Blum and Ingold2020).
A link to the inaugural issue
Even though it only dates back roughly 15 years, it is a testament to research- and practice-based interest in ACs that the first issue of Industrial-Organizational Psychology: Perspectives on Science and Practice (IOP) included a focal article on ACs by Charles Lance along with responses to his article.Footnote 1 Lance (Reference Lance2008) contributed a critique of how ratings from ACs are scored. He concluded that ACs do not measure dimensions and that attempts to use ACs to generate dimension scores should be abandoned. Points for consideration, reactions, and diverging points of view raised by Lance’s critique (e.g., Arthur et al., Reference Arthur, Day and Woehr2008; Howard, Reference Howard2008; Rupp et al., Reference Rupp, Thornton and Gibbons2008) have consistently elicited novel research on ACs.
During the 15 years that have passed since the Lance (Reference Lance2008) focal article and its responses, perspectives on ACs have continued to develop and have benefitted from knowledge generated from ongoing research. This special issue on ACs provides an opportunity to reflect again on what conceptual perspectives on ACs have prevailed and emerged since 2008 as well as an opportunity to explore and showcase recent insightful empirical AC research. It moreover provides insights of value to practice and on how ACs can continue to provide an abundant source of information for organizational research and for decision makers.
Overview of papers in the special issue
Dewberry (Reference Dewberry2024) reviews research literature concerned with whether ACs assess dimensions (or competencies, e.g., communication skills, tolerance) reliably and as intended. Fifteen years after Lance’s focal article, Dewberry reviews more recent research on whether ACs measure dimensions, particularly research utilizing generalizability theory (G theory), which provides statistically controlled estimates of dimension effects. He concludes that evidence derived from G theory research confirms that ACs do not measure dimensions and concurs with Lance that attempts to measure dimensions with ACs should be abandoned. Dewberry moreover presents an argument against interactionist perspectives on ACs (e.g., trait activation theory, the mixed-model perspective), suggesting that some of the patterns used to support these perspectives may simply reflect artifacts of the AC measurement design. This work raises considerations for practitioners about whether they should cease attempts to use ACs to measure dimensions, scoring them instead only in relation to exercises and/or overall performance.
In a conceptual paper, Nottingham and Rupp (Reference Nottingham and Rupp2024) propose that ACs could be used to serve the emerging aim of selecting and developing inclusive leaders in organizations. In contrast to the Dewberry paper, using a dimension approach, Nottingham and Rupp argue that measuring inclusive leadership may add incremental validity to overall assessment ratings (OARs). Specifically, Nottingham and Rupp suggest assessing inclusive leadership proficiency as a behavioral leadership dimension and they develop specific propositions about its relationship with leadership and follower performance and diversity among followers. Given that the evaluation of leaders and the fostering of diversity in organizations are key considerations for both research and practice, this article provides valuable guidance on how to potentially optimize evaluation criteria when aiming to assess inclusive leadership.
Prior research has illustrated that AC ratings are dependent on impressions that assessors form of candidates (Ingold et al., Reference Ingold, Dönni and Lievens2018; Lance et al., Reference Lance, Foster, Gentry and Thoresen2004). Yet it is unknown to what degree these impressions remain consistent in ACs and if the impact of these impressions on AC ratings changes across and within AC exercises. Building on the thin-slice paradigm, Ingold et al. (Reference Ingold, Heimann and Breilin press) address this topic and investigate the consistency of assessor candidate impressions by using different slices of video material from the beginning, middle, and end of three AC exercises. Results suggest that the impressions participants convey across different time points of the AC (i.e., the beginning, middle, and end of each AC, but also across exercises) are consistent. Moreover, their findings suggest that these impressions predict AC performance and can also relate to participant’s job performance. This study offers research insights into the relevance of assessor impressions and into the development of our understanding of assessee behavior.
The topic of assessor training for ACs, which can be positioned in the broad research area of frame-of-reference training (Roch et al., Reference Roch, Woehr, Mishra and Kieszczynska2012; Woehr & Huffcutt, Reference Woehr and Huffcutt1994), is addressed by Gorman et al. (Reference Gorman, Jackson, Meriac and Himmler2024). Applied to ACs, frame-of-reference training provides assessors with a common set of standards to evaluate performance with the goal of increasing rater effectiveness and consistency. In their study, Gorman et al. provide a perspective on the multifaceted structure of frame-of-reference training. The authors found that assessor training was of most assistance in the identification of low-performing AC participants. They moreover found that ratings for assessors who were not trained were associated with larger proportions of residual error than for those who were trained. This study contributes knowledge to research and practice associated with AC training and provides insights into how the variance profile of AC ratings depends on whether assessors have been trained.
In her article, Roch (Reference Roch2024) provides evidence for perceptual differences among applicants relating to AC exercises and an ability test. Different applicant perceptions were found for different exercise types. Whether the AC was rated live or via a recording had implications for fairness perceptions. Moreover, Roch found that whether an assessee had previous experience on an AC influenced levels of perceived self-efficacy. This study contributes to knowledge on applicant reactions that can influence a practitioners’ choice of exercises and psychometric tests.
Procedural justice is conceptually related to considerations of ethics, which is a topic of major consequence to organizations. Fostering just, moral, and ethical behavior is paramount: not only for organizations but also for the wider development of society. d’Amato et al. (Reference d’Amato, Murugavel, Mereiros and Watts2024) address this issue in their paper and raise questions about how leaders can develop ethical and moral behavior using the AC method. They provide initial findings suggesting that the development of ethical leadership attitudes may result in negative, backlash-oriented repercussions. For research, this study offers insights into the application of ACs in the context of ethical leadership. For practice, it provides early warnings about some of the pitfalls of attempting to develop attitudes with ACs.
Organizational decision making based on AC results is complex and, as Rupp et al. (Reference Rupp, Thornton, Bisbey, Nottingham, Salas and Murphy2024) suggest, it can involve the decision maker in a consideration of theory, empirical contributions, and best-practice considerations. In their paper, Rupp et al. present an epistemology for the integration of these three factors. They apply their framework to present a perspective on assessment and development that is directly relevant to considerations of ACs. They conclude that there are areas of alignment among theory, empirical contributions, and best practice. However, they also highlight key gaps and areas for further development. The Rupp et al. framework could be applied by both researchers and practitioners to assist in furthering research to help ensure a better integration across theory, research, and practice in assessment and development and also in other areas of complex workplace interventions.
Conclusions
This special issue offers an exploration into existing, new, and alternative lines of inquiry that showcases the progressive, enterprising, and current nature of AC research and development. In our view, modern ACs possess a vast wealth of knowledge and value to offer both individuals and organizations and will continue to stimulate research as they have before and following the first issue of IOP. In our reading, the current collection of works suggests that researchers and practitioners are served best by exploring, debating, and engaging with areas for development, challenges, and controversies associated with ACs. We propose that learning more about such issues, stimulating debate around them, and allowing different perspectives to be heard is how a research area can grow and develop. We hope that the current set of papers will contribute to the AC debate, will reignite unresolved controversies, and will stimulate new lines of enquiry. It is through such debates and discussions that we can, as researchers and practitioners, further strive toward an enhanced understanding of the AC method to the continued benefit of individuals and organizations.
Competing interests
None.