Professional Documents
Culture Documents
BARTRAM Original ASSESSMENT Articles IN ORGANISATIONS Blackwell Oxford, Applied APPS 0269-994X April 0 1 2 53 00 International 2004 UK Psychology: Publishing Association an LtdInternational for Applied Review Psychology, 2004
Assessment in Organisations
Dave Bartram*
SHL Group plc, UK
Cet article aborde les pratiques actuelles et les tendances mergentes de lvaluation dans les organisations. Une attention particulire est accorde a lvaluation en vue du recrutement et de la selection, l o lapparition de la mta-analyse a fondamentalement chang la conception que lon pouvait avoir des tests psychologiques et autres techniques de slection. On analyse aussi limpact dInternet sur les pratiques de slection. En ce qui concerne lvaluation postembauche, lobligation pour les organisations dassumer des changements rapides est raporte limportance de la modlisation des comptences. Quelques points-cls de lvaluation post-embauche sont passs en revue (leadership, feedback 360). Des perspectives pour la nouvelles recherches sont esquisses; il sagit de crer des thories et des modles plus pertinents, mais aussi de progresser dans les tudes de validit partir les donnes existantes. Finalement, on saperoit que lessentiel de la littrature actuelle sappuie sur des recherches ralises aux Etats-Unis, un peu au Royaume Uni ou dans dautres pays dEurope. Beaucoup de ces recherches sont limites dans leurs possibilits dapplication en ce sens quelles ont t mences sur de grandes organisations. On insiste sur le manque de travaux interculturels et sur la ncessit de sintresser lensemble des organisations de travail (des grandes aux petites, des enterprises locales aux multinationals, aux secteurs public et priv). The article considers current practice and merging trends in assessment in organisations. Particular attention is paid to assessment for recruitment and selection, where the use of meta-analysis techniques has radically changed the way in which psychological tests and other selection techniques are viewed. The impact of the Internet on selection practice is also discussed. For posthire assessment, the impact of the need for organisations to undergo rapid change is considered in relation to the importance of competency modelling. Some key areas (leadership, 360-degree feedback) of post-hire assessment are reviewed. Issues for future research are outlined. These include the need for better theory and models, together with the need to move ahead of a reliance on old data sets. Finally, it is noted that much of the current literature is based on research in the United States (with some from the United Kingdom and other parts of Europe). Much of the research is also limited in applicability in that it is based on large organisations. The need for more cross-cultural studies
* Address for correspondence: Dave Bartram, SHL Group plc, The Pavilion, 1 Atwell Place, Thames Ditton, Surrey KT7 0NE, UK. Email: Dave.Bartram@shlgroup.com International Association for Applied Psychology, 2004. Published by Blackwell Publishing, 9600 Garsington Road, Oxford OX4 2DQ, UK and 350 Main Street, Malden, MA 02148, USA.
238
BARTRAM
and the need to cover the full range of work organisations (large to small; local to global; private to public sector) is emphasised.
INTRODUCTION
Assessment is carried out by organisations as a means of measuring the potential and actual performance of their current (post-hire assessment) and potential future employees (pre-hire assessment). Measurement is important because it enables organisations to act both tactically and strategically to increase their effectiveness. Pre- and post-hire assessment practices are typically referred to as assessment for selection and recruitment, and assessment for development and performance management, respectively. While organisations make considerable use of objective assessment techniques for both, by far the most extensive literature is on the validity of assessment methods for recruitment and selection. Most research on post-hire interventions tends to focus on the efcacy of training or development interventions rather than the assessments associated with them. The main area of post-hire assessment where instrumentation has been researched in some detail is that of management and leadership assessment. In most other areas, there has been a tendency for organisations to be far less concerned about the quality and validity of the tools they use for post-hire assessment (e.g. for 360-degree feedback) than they are about those they use for selection. The present paper will consider both pre- and post-hire assessment. While the main focus will be on the former, aspects of post-hire assessment will also be addressed. In particular the paper addresses the issue of the need for a generalisable conceptual framework for assessing behaviour in the workplace.
ASSESSMENT IN ORGANISATIONS
239
240
BARTRAM
As personality measures tend to be independent of measures of ability, they can add signicant increments to the overall validity of any selection battery. The most powerful combination, on the basis of current evidence (Ones, Viswesveran, & Schmidt, 1993; Ones & Viswesveran, 1998), is that provided by ability tests and measures of integrity or conscientiousness. Together these provide (adjusted) validities around .65. Ones and Viswesveran (2001) have shown that, paradoxically, while measures of integrity provide better prediction of job performance than honesty criteria, the reverse is the case for Conscientiousness. A possible explanation for this is that Conscientiousness is actually a narrower measure than that provided by integrity tests. The latter may well provide an assessment of what Digman (1997) has termed the Alpha Factor, which brings together Agreeableness, Conscientiousness and Emotional Stability. Digman (p. 1249) argued that factor alpha represents the socialization process itself . . . concerned with the development of impulse restraint and conscience, and the reduction of hostility, aggression, and neurotic defense. Finally, personality measures have the advantage of reducing adverse impact. A major concern over the use of ability tests is that they tend to show large group differences, with blacks scoring from .5 to 1.0 SD lower on average than whites (Roth, Bevier, Bobko, Switzer, & Tyler, 2001). Such differences can result in adverse impact (i.e. the selection of a disproportionate number from one group relative to another) if applicants are selected on the basis of ability alone. Research has shown, however, that when ability and personality measures are combined, the group differences on composite score measures are greatly reduced, generally to a level where adverse impact is no longer a problem (Baron & Miles, 2002). Though the above discussion has focused on ability and personality testing, meta-analysis has also claried issues of validity in relation to a range of other selection tools. For example, it is now clear that the interview, once dismissed by academic psychologists as having no validity, can have good validity if it is appropriately structured and criterion referenced. While meta-analysis has revolutionised thinking in the area of personnel selection, it is important to keep in mind a number of caveats.
International Association for Applied Psychology, 2004.
ASSESSMENT IN ORGANISATIONS
241
First, meta-analysis outcomes are dependent on the data sets used. It is probably true to say that, in the past ten years, we have now exhausted the information that can be drawn from the body of studies that exists in the literature. Second, meta-analyses are historical. That is, they draw general conclusions from historical bodies of data. These data sets were collected, by and large, at a time when measures of criterion performance were weak, and when the formal use of instruments like personality questionnaires was not encouraged. Hence there is a danger of using their results to shape current and future practice in a way that holds back progress if one is not careful. Third, there is a danger that people will rely on the results of metaanalyses as somehow dening the validity of procedures, and regard it as unnecessary to do further validity studies. Finally, the data sets used in these analyses are drawn predominantly from the United States. While the general pattern of results has been supported by some more recent analyses (Salgado, 1998; Salgado, Anderson, Moscoso, Bertua, de Fruyt, & Rolland, 2003) that have looked at European data sets, it is important to consider how selection and recruitment practices differ around the world.
International Variations
With globalisation, international companies have had to try to decide how to implement assessment practices on a worldwide scale. Local companies are also now competing in their national markets with global multinationals. As such, they are under pressure to adopt the practices of the big players. Employee selection is probably the area where both formal and informal assessment methods are most intensively used by organisations. It is also the area that has most exposure, in that organisations are assessing people who are external to the organisation. As a consequence, the assessment practices adopted are very public and will affect the image of the company. They can have either positive or negative public relations value. It is of interest to study differences in assessment practices between countries for two reasons. First, it provides a potential basis for comparing the effectiveness of alternative approaches to assessment and provides the basis for practitioners in each country to reect on the value of the practices they use. Second, it is becoming increasingly important because of the globalisation of industry and the world of work. International organisations are now seeking to adopt common selection and recruitment practices across different countries. Organisations who are seeking to grow their business in new markets will be concerned both to recruit the best local managers and workers in that market and to identify people on their current staff who
International Association for Applied Psychology, 2004.
242
BARTRAM
will operate effectively as expatriates to kick-start growth in such markets. Should they adopt local practices in selection assessment or should they impose their head ofce practices on this new environment? Various national and international surveys of the use of these methods were published in the 1990s (Shackleton & Newell, 1991, 1994; Rowe, Williams, & Day, 1994; DiMilia, Smith, & Brown, 1994; Gowing & Slivinski, 1994; Funke, 1996; Levy-Leboyer, 1994; Bartram, Lindley, Foster, & Marshall, 1995; Ryan, McFarland, Baron, & Page, 1999). Unfortunately most of these studies tend to cover only a single country or group of countries in the developed world (typically, North America, Europe, and Australia) and most rely on postal surveys with low return rates. Levy-Leboyer (1994), in her review of surveys of selection practices in Europe, noted that all countries tend to use application forms and interviews. France is unusual in placing a high reliance on graphology as a screening method and in not giving much credence to third-party references. Situational tests and assessment centres are used more in the UK, Germany, and the Netherlands than in France or Belgium, but there are also generally higher levels of test use in France, Belgium, Spain, and the UK than in the other countries. Schuler et al. (in a 1993 study reported in Funke, 1996) found very high levels of use of structured interviews in the Benelux countries and the UK, with less in Germany. In Germany they found very little use of personality or performance testing compared to Spain, Benelux, and the UKwith France in-between. The use of objective assessment techniques (tests, structured interviewing, assessment centres) raises issues of training. Many authors, for example, have expressed concerns about the indiscriminate use of tests in countries where there is little if any provision for training in these methods (e.g. OGorman, 1996, with reference to Australia; Engelhart, 1996, with reference to France; Henley & Bawtree, 1996, with reference to the UK; and Smith & George, 1992, with reference to New Zealand). It is in response to these concerns that the British Psychological Society established the procedures for Certication of Competence in Testing for people working in occupational testing (Bartram, 1995). More recently the International Test Commission has published international guidelines for test use (Bartram, 2001; International Test Commission, 2001). The various surveys listed above are interesting in that they tell us something about how countries differ. But they are limited in that they almost all only concern large organisations (1,000 or more employees), they rely on postal surveys that often have very low response rates, and they say nothing about why these differences in practice occur. An exception to the postal review methodology is a study reported by Bartram et al. (1995). They carried out detailed face-to-face interviews with recruiters using a structured sampling technique covering small businesses
International Association for Applied Psychology, 2004.
ASSESSMENT IN ORGANISATIONS
243
in the UK. This research focused on the question of how young people are selected for employment by small businesses. It is often forgotten that the majority of people work for small companies. In the UK at the time of this research, over 88 per cent of businesses employed fewer than 25 people and 73 per cent employed fewer than ten. About one-third of all employed people work for small businesses, yet very little attention has been paid to their recruitment and assessment practices. The research conrmed the prediction that small rms would carry out assessment in a very casual and informal fashion. The impression gained from the psychological literature, that assessment in industry is objective and well formalised, is a misleading one. Actual practice differs considerably from what assessment specialists would dene as best practice. It is also interesting to consider what recruiters assess and their perceptions of the worth of various assessment methods. The Bartram et al. (1995) study and subsequent work (Coyne & Bartram, 2000) has shown that the most important personal characteristics for employers are: honesty, integrity, conscientiousness, interest in the job, and the right general personality. Research by Scholarios and Lockyer (1999) with small consultancy rms in Scotland again found an emphasis on the importance of honesty and conscientiousness, with general ability as the third most important attribute. These psychological attributes are regarded as more important than qualications, experience, or training. One of the main reasons given by employers in the Bartram et al. (1995) study for their reliance on the interview was that they thought this was the best way for them to judge an applicants personal qualities. This emphasis on the importance of personal qualities probably accounts for the growing use of personality tests, measures of emotional intelligence and honesty and integrity testing by mediumsized and larger organisations together with the relative lack of emphasis given to formal educational and work-related qualications (Jenkins, 2001). Why do recruiters focus on these particular characteristics? Evidence suggests that these are characteristics that they see as relatively difcult to change and high risk areas. A person can be trained to acquire the skills and knowledge necessary to do a job, but their attitude, their way of dealing with other people, their honesty, are seen as characteristics that are immutable. This is highlighted in small rms where the impact of one person on the performance of the business as a whole can be very considerable. Campbell, Lockyer, and Scholarios (1997), who studied medium- to small-sized rms (most of their sample employed fewer than 200 people), also found that such organisations tended not to use objective assessment techniques (such as psychological tests) and also tended not to use assessment centres (because of their cost). While Boyle, Fullerton, and Yapp (1993) show that use of assessment centres is mainly conned to large rms, there is some evidence from an
International Association for Applied Psychology, 2004.
244
BARTRAM
Institute for Recruitment Studies survey (IRS, 1997), that medium-sized rms are making increasing use of these techniques. In the most extensive of the published surveys of international selection practices, Ryan et al. (1999) attempted both to increase the breadth of coverage of practices around the world and to explore some possible explanations of why they differ. Through a postal survey, they mailed 300 organisations employing more than 1,000 people in each of 22 countries (i.e. n = 6,600 in total), including Singapore, Hong Kong, Malaysia, and South Africa. A total of 959 usable responses were obtained. Cultural norms for each country on Uncertainty Avoidance and Power Distance were obtained from Hofstedes (1991) data. The Ryan et al. study is worth considering in some detail as it directly focuses on the issues of how practices differ between countries and whether such differences can be related to differences in cultural norms. They found that between-nation effects could explain a modest proportion of variance in stafng practices. In particular, they found that national differences accounted for a substantial proportion of the variance in the use of xed interview questions (43% of the variances), and in using multiple methods of verication, testing, and number of interviews (over 10% of the variance in each case). In relation to the predictions based on Hofstede, some interesting results were found. Those cultures high in Uncertainty Avoidance tended to use fewer selection methods, do rather less verication and background checking, but use more types of tests and use them to a greater extent, and conduct more interviews with candidates. They were also more likely to use xed sets of interview questions and more likely to audit their selection procedures in some manner. Less clear results were found in relation to expected effects of Power Distance differences. While countries high on this scale were less likely to use peers as interviewers, peers were more likely to be involved in hiring decisions in some other way (possibly because of the impact of unionised labour in such countries). Future research needs to tackle the difcult questions of sampling. Ryan et al. had very different response rates from different countries. There were also geographical gaps: South America, East and West Africa, the Middle East, and North Africa. Arguably these are the areas where we would expect to nd the greatest cultural differences and are key emerging markets for multinationals. It is also possible that other cultural frameworks would have provided different views of the data (e.g. Schwartz & Bilsky, 1987; Smith, Dugan, & Trompenaars, 1996). Nevertheless, this study provides a valuable starting point for future research in this area. What the various research studies on recruitment practice have shown is that there is wide diversity in the use of assessment methods in organisations both within and between countries. This diversity is very much a function
International Association for Applied Psychology, 2004.
ASSESSMENT IN ORGANISATIONS
245
of both organisation size and culture. Small organisations have neither the resources nor the people with the necessary specialist skill and knowledge to devote to testing and assessment. When comparing practices in different countries, it is important that we do not compare the practices of small rms in developing countries with those of large ones in developed countries. Small rms in the UK, for example, rely heavily on personal recommendation and family and friendship network connections for nding new employees. This sort of practice is often, wrongly, seen as occurring only outside the North American and European economic areas. Ryan et al.s study is important for showing clearly that national and cultural factors do account for substantive proportions of variance in difference in selection practices. Such differences raise questions for international organisations seeking to carry out fair recruitment campaigns across a number of different countries. Should they impose the same practices on all countries? If they do, then these practices will be more familiar in some countries than others, and more acceptable (both to applicants and recruiters) in some than others. The alternative is to set quality standards for the recruitment process that do not prescribe methods as such, and leave each country to adopt the methods it is most comfortable with so long as the outcome is one that meets the quality standard. Despite the variations in assessment practice, there does appear to be good agreement across organisations and cultures on the importance of personal qualities, on the importance of a persons values, and the degree to which a new recruit will t the culture of the organisation.
Competency-Based Assessment
The competency approach to selection and assessment is one based on identifying, dening, and measuring individual differences in terms of specic work-related constructs that are relevant to successful job performance. Over the last 25 years this approach has gained rapidly in popularity, due partly to the way in which the concepts and language used have currency within the world of human resources management. The proling of jobs in terms of competency requirements has increasingly supplemented or replaced more traditional task-based job analysis, most noticeably in countries outside the United States. Competency proling differs from job analysis in that the focus of the former is on the desirable and essential behaviours required to perform a job, while the latter focuses on the tasks, roles, and responsibilities associated with a job. These are complementary ways of looking at the same thing, with the competency analysis providing a person specication and the job analysis a job description. The
International Association for Applied Psychology, 2004.
246
BARTRAM
main advantage of the competency modelling approach has been its success in building the models that lay the foundations for organisation-wide integrated human resources applications. The problem with competency as a construct is that there is considerable confusion and disagreement about what competencies are and how they should be measured (Shippmann, Ash, Battista, Carr, Eyde, Hesketh, Kehoe, Pearlman, Prien, & Sanchez, 2000). Competency-based assessment has also suffered in the past from being used and developed by a wide range of practitioners many of whom had not had a psychologists background of training in scientic method and measurement. However, Shippmann et al. (2000) note that there is evidence of increasing rigour in the competency approach. Bartram, Robertson, and Callinan (2002) dene competencies as sets of behaviours that are instrumental in the delivery of desired results or outcomes. In terms of this denition, competencies relate to behavioural repertoires: the range and variety of behaviours we can perform, and outcomes we can achieve. A competency is not the behaviour or performance itself but the repertoire of capabilities, activities, processes, and responses available that enable a range of work demands to be met more effectively by some people than by others. This approach is elaborated in Kurz and Bartram (2002).
ASSESSMENT IN ORGANISATIONS
247
results of subject matter experts sorting 147 behavioural elements. The nine general areas were: Traditional functions; Task orientation; Dependability; Open-mindedness; Emotional control; Communication; Developing self and others; Occupational acumen and Concerns. Borman and Brush (1993) propose a structure of 187 behaviours mapping on to 18 main dimensions, which in turn map to four very broad dimensions: Leadership and supervision; Interpersonal relations and communication; technical behaviours and mechanics of management; and useful behaviours and skills (such as job dedication). This structure has been supported by subsequent meta-analysis research (Conway, 1999). Kurz and Bartram (2002) describe a job competency framework which also adopts a three-tier structure. The bottom tier of the structure consists of a set of 112 component competencies. These 112 component competencies were derived from extensive content analyses of both academic and practice-based competency models. This analysis covered managerial and non-managerial positions. As a consequence, the content of the components covers a wider domain than that addressed by Tett et al. in their work on managerial competencies. The framework articulates the relationships between these components, their mapping on to a set of 20 competency dimensions (the middle tier) and their loadings on eight broad competency factors (the top tier in their Table 1). These models provide the basis for the development of the measures of behaviour at work that are needed if we are: To assess the general value of trait measures in applications like personal development, potential for leadership, assessment for promotion and so on. To get a better-articulated model of criterion measures for the validation of selection and recruitment procedures. In the past, there has been a tendency to use tools in selection (and for posthire assessment) that were designed to provide coverage of some particular psychological attribute domain (e.g. personality) rather than having been designed to provide coverage of the relevant criterion domain. A better understanding of the factorial structure of the domain of criterion behaviours will help us to better design predictors both in terms of coverage and validity. As noted earlier in this article, two factors came together to create the shift in attitude relating to the validity of personality assessment. One was the availability of meta-analysis techniques. The other was the Big Five taxonomy. This allowed various different instruments to be mapped on to
International Association for Applied Psychology, 2004.
248
BARTRAM
a common structure for the purposes of analysis, and issues of generalisability to be explored across different instrument structures. A similar issue arises in relation to assessing behaviours in the workplace. So long as there is nothing comparable to the Big Five for this domain, we will continue to face the problem of trying to aggregate the results from studies that use criterion measures that are not comparable with each other. However, evidence is beginning to emerge that there may be a small number of broad factors that account for most of the variance in criterion workplace behaviors. The research reported in Kurz and Bartram (2002) supports the view that variance in competency measures can be accounted for by eight broad factors. Correlations between competency measures and measures of psychological attributes suggest that these eight factors reect the psychological constructs that underlie competencies. Specically the trait markers for the eight competency factors can be identied as: g or general reasoning ability the Big Five personality factors two factors relating to motivation factors: need for achievement and need for power or control. This Great Eight structure has been replicated in a number of different data sets including analysis of the ratings of 54 competencies in the OPQ32 UK national standardisation sample data (SHL, 1999), analysis of job applicant data collected over the Internet in the USA (n = 26,000), from Swedish data, and from analyses of data obtained from two generic 360-degree competency inventories. This model suggests that optimal selection batteries will be those that provide coverage of general ability, the Big Five personality factors and two broad motivation factors. Meta-analyses of selection data have already supported the importance of the ability and personality domains. The fact that motivation factors have not emerged from meta-analyses is a reection of the fact that there are very few studies in the literature that contain systematic measures of motivation. We should now be designing new validity studies that start from a consideration of the criterion domain rather than from the availability of classes of predictor instrument. Assessment in organisations tends to be focused around competency models. The use of personality questionnaires and other trait-based measures is generally carried out within the context of assessing competency potential rather than for any direct interest in the traits themselves. Trait measures tend to be used in Development Centres as part of the process of exploring peoples potential for future development. More generally, assessment of behaviour and of performance is more direct; relying on instruments designed to measure samples of behaviour rather than signs of underlying traits.
International Association for Applied Psychology, 2004.
ASSESSMENT IN ORGANISATIONS
249
250
BARTRAM
some cultures and impediments in others. However, the GLOBE project has only asked about the perceived desirability of behaviours. It is an open question as to whether such behaviours are also effective in facilitating organisationally desired outcomes.
EMERGING TRENDS
From what has been said earlier, it is clear that a number of factors are impacting the role of assessment in and by organisations. These can be divided into some general issues and others which are more specically associated with pre- or post-hire assessment.
General Issues
Meta-analysis and validity generalisation have had a major impact on thinking and practice in the eld. The emphasis on small local validation studies has been replaced by a focus on the need to consider aggregated data sets, where various sources of small-sample error have been taken into account. The strengths and weaknesses of this approach have been discussed (see also, Murphy, 2000). A current limitation of this literature is its reliance on old data sets. These do not reect current thinking on the structure and breadth of the criterion space. It will be some time before there are sufcient well-designed studies completed which provide the range of prediction measures needed to fully evaluate the potential upper limits on validity for assessment in selection. A second general issue is that associated with the growth in the use of objective assessment in developing countries. Multinationals operate worldwide; increasingly they are looking to use the same techniques in developing nations for selection and development as are used in the so-called developed world (tests, Assessment and Development Centres, structured interviews, biodata inventories and so on). These raise issues of local user competence as well as cultural appropriateness. One consequence of this trend has been the development of black box solutions. Such solutions seek to de-skill the user requirements by providing a computer-based expert system that guides use through the process of job analysis and competency proling; selects and congures a battery of relevant tests and provides the user with a merit list of applicants, sorted by their performance on the tests in relation to the job requirements. Such systems will also, typically, provide the user with reports that give an interview structure and identify key areas for questioning. While such systems put the knowledge and skill of the expert I /O psychologist into the hands of line managers, concerns have been raised about the impact of this on the future role of I/O psychologists in assessment in organisations.
International Association for Applied Psychology, 2004.
ASSESSMENT IN ORGANISATIONS
251
Pre-Hire Issues
This review has shown that not only do employers value honesty and conscientiousness above all other attributes, but also that meta-analyses have shown that these attributes are well measured by integrity tests and personality questionnaires. More and more organisations want to assess employee dependability, especially in blue-collar and lower level whitecollar positions. While research tends to show that both covert and overt measures of honesty do have generalisable validity, there is a range of ethical and legal concerns about their use. Outside the United States, professional concerns are expressed about the use of overt honesty measures, while less concern is shown over the use of personality attributes like conscientiousness. Within the United States, the opposite appears to be the situation, where overt questions are preferred to what would be regarded as the covert use of personality to infer honesty. As noted earlier, it is paradoxical that the measures of conscientiousness actually appear to assess honesty better than general job performance while the reverse is the case for integrity tests. Future research needs to better understand the mechanisms behind these patterns of prediction. The ethical issues would seem to be more about how these instruments are used than whether they should be usedso long as their job relevance can be supported. Probably the biggest change in recruitment and selection practice in the past few years has been brought about by the use of the Internet. Use of the Internet as the medium for job search and making job applications is rapidly replacing the traditional paper-based application procedures. Details of how this technology has impacted on assessment practice are reviewed in Bartram (2000), Lievens and Harris (2003), Robertson, Bartram, and Callinan (2002), and Stanton (1999). From the assessment point of view, one impact has been the pressure on test developers to design fast, open access instruments that still meet good psychometric measurement criteria. This has resulted in the emergence of four key research areas. In order to ensure better test security and provide the ability to refresh test content on a regular basis, there has been a rapid increase in the application of item and test generation methodologies (Irvine & Kyllonen, 2002). A practical example of this is provided by Baron, Miles, and Bartram (2001). The increasing use of computerised testing has resulted in research on the equivalence of computer and paper-and-pencil tests. In general, this has produced generally positive results indicating that tests, especially un-timed self-report measures, need not be affected by this change of medium (Bartram, 1994; Donovan, Drasgow, & Probst, 2000; Neuman & Baydoun, 1998). However, highly speeded ability tests (such as clerical checking tasks) may need re-norming as the ergonomics
International Association for Applied Psychology, 2004.
252
BARTRAM
of the test can signicantly affect the speed with which people can perform the task. While organisations are looking for means of increasing test security and protecting tests from compromise through coaching, at the same time they are wanting to make assessment procedures more open and less constrained by the need to provide supervision, and more accessible to line managers or hiring managers with little or no expertise in testing. This has increased the need for research on the impact supervision has on performance in assessment situations, on how one can maintain validity under conditions of reduced control, and on how effectively selection solutions can be black-boxed for use by relatively inexperienced line managers. Finally, there is increased interest in the search for valid predictors that minimise adverse impact. Organisations are increasingly concerned about potential litigation and the socio-political implications of selection practices. In the past, attention tends to have focused on the problems associated with the use of ability tests, rather than the overall selection assessment strategy. We are now seeing how, with the judicious use of combinations of instruments, the bottom line risk of adverse impact can be greatly reduced. While meta-analyses might imply that the only tools one needs are an ability test and an integrity test and that these could be used for all jobs without any job analyses being needed, the reality is different. As argued earlier, meta-analysis has provided some important breakthroughs, but it has its limitations. Selection is a process of negotiation between applicants and hirers. Assessment techniques have not only to work, but also to appear relevant and acceptable. In a study comparing paper-and-pencil, computerised, and multi-media versions of a test (Richman-Hirsch, Olson-Buchanan, & Drasgow, 2000), managers rated the multi-media version as more valid and had more positive attitudes towards it than did those completing the other versions. Assessment procedures are seen as reections of an organisations values and culture. The choice of which tools to use and how to apply them is governed by more than just considerations of psychometric quality.
ASSESSMENT IN ORGANISATIONS
253
who have the qualities necessary to work exibly and adaptively within an organisation. This in turn has pushed the emphasis away from task-based job analysis as the basis for developing person specications, towards the use of competency modelling. Competency models provide the means whereby an organisation can integrate the tactical and strategic use of assessment. As discussed earlier, there is a growing consensus on the nature and structure of the competency domain. The Great Eight factors provide a useful starting point for relating assessments of traits to measures of workplace behaviours. This emphasis on the need for a exible workforce has increased the use of assessment for development purposes within organisations. Approaches such as 360-degree feedback are becoming increasingly common. The logistical difculties of carrying out a 360-degree assessment have been largely overcome by the development of Internet-based solutions (Bartram, Geake, & Gray, in press). In contrast to pre-hire assessment, post-hire approaches are characterised by the use of multiple sources: most leadership assessment instruments, for example, involve both self-ratings and peer or subordinate ratings. We are also seeing an increasing use of assessment instruments for organisational development (OD) applications: Identifying organisational culture and values, looking at value t between individuals, groups, and the organisation and so on. These raise issues of the impact of cultural and national differences between organisations, which are being addressed using models developed for cross-cultural analysis (e.g. Hofstede, 1991; Smith et al., 1996; Schwartz, 1999).
254
BARTRAM
provides a way forward in developing a generalisable structure of factors that can be used to measure workplace behaviours, and so provide a common structure for future validation studies. In so doing, it is important to draw a distinction between two categories of criterion: workplace behaviours (as addressed by competency models) and outcomes (performance judged against the achievement of goals or objectives). Schneider (1996), reviewing a series of papers on personality and work, noted that:
. . . knowing that Conscientiousness correlates with various performance criteria across a wide variety of jobs and in a wide variety of settings is not equivalent with understanding the behaviour that is emitted by those who are conscientious. Campbell (1990) is correct when he notes that we have focused on outcomes of behaviour as correlates of personality (and other predictors) and have relatively little insight into the behaviour that intervenes between the personality and the outcome. In the absence of such information, we have no understanding of the process by which personality becomes reected in outcomes. This lack of information leaves us relatively impotent so far as interventions in the workplace are concerned. ( pp. 291292)
He goes on to point out that It is behaviour, not personality that causes outcomes. This important distinction has been missed in much research on assessment in organisations over the past decade. Future research needs to be guided by better models and by theory. We should be seeking to understand the dynamics of the processes that relate attributes to behaviours, and how behaviours can be shaped to produce outcomes that meet organisational goals. Without demeaning the very real contribution made by the metaanalytic literature, we need to be less driven by the pure empiricism that can characterise such approaches and the concomitant recycling of data from the past, and focus more on building a sound theoretical base from which to model the processes that lead from individual behaviours to organizational effectiveness.
REFERENCES
Alban-Metcalfe, R.J., & Alimo-Metcalfe, B. (2002). The development of the Transformational Leadership QuestionnairePrivate Sector Version. Proceedings of the British Psychological Society. Leicester: BPS. Alimo-Metcalfe, B., & Alban-Metcalfe, R.J. (2001). The construction of a new Transformational Leadership Questionnaire. Journal of Occupational and Organizational Psychology, 74, 1 27. Baron, H., & Miles, A. (2002). Personality questionnaires: Ethnic trends and selection. Proceedings of the British Psychological Society Occupational Psychology Conference ( pp. 110 116). Leicester: BPS.
International Association for Applied Psychology, 2004.
ASSESSMENT IN ORGANISATIONS
255
Baron, H., Miles, A., & Bartram, D. (2001). Using online testing to reduce timeto-hire. Paper presented at 16th Annual Conference of the Society for Industrial and Organizational Psychology, San Diego, April. Barrick, M.R., & Mount, M.K. (1991). The Big Five personality dimensions and job performance: A meta-analysis. Personnel Psychology, 44, 125. Barrick, M.R., Mount, M.K., & Judge, T.A. (2001). Personality and performance at the beginning of the new millennium: What do we know and where do we go next? International Journal of Selection and Assessment, 9, 930. Bartram, D. (1994). Computer-based assessment. International Review of Industrial and Organizational Psychology, 9, 31 69. Bartram, D. (1995). The development of standards for the use of psychological tests: The competence approach. The Psychologist, 8, 219223. Bartram, D. (2000). Internet recruitment and selection: Kissing frogs to nd princes. International Journal of Selection and Assessment, 8, 261274. Bartram, D. (2001). The development of international guidelines on test use: The International Test Commission Project. International Journal of Testing, 1, 33 53. Bartram, D., Geake, A., & Gray, A. (in press). The Internet and 360-degree feedback. In S. Martin (Ed.), 360-grad-beurteilungen, diagnose und entwicklung von fuhrungskompetenzen. Gottingen: Verlag fur Angewandte Psychologie Hogrefe. (Reihe Psychologie fur das Personalmanagement; Herausgeber: Prof. Dr Werner Sarges.) Bartram, D., Lindley, P.A., Foster, J., & Marshall, L. (1995). The selection of young people by small businesses. British Journal of Occupational and Organizational Psychology, 68, 339 358. Bartram, D., Roberton, I.T., & Callinan, M. (2002). Introduction: A framework for examining organisational effectiveness. In I.T. Robertson, M. Callinan, & D. Bartram (Eds.), Organisational effectiveness: The role of psychology (pp. 112). Chichester, UK: Wiley. Bass, B.M. (1985). Leadership and performance beyond expectations. New York: The Free Press. Bass, B.M., & Avolio, B.J. (1990a). Multifactor Leadership Questionnaire. Palo Alto, CA: Consulting Psychologists Press. Bass, B.M., & Avolio, B.J. (1990b). Transformational leadership development: Manual for the Multifactor Leadership Questionnaire. Palo Alto, CA: Consulting Psychologists Press. Borman, W.C., & Brush, D.H. (1993). More progress towards a taxonomy of managerial performance requirements. Human Performance, 6, 121. Borman, W.C., Penner, L.A., Allen, T.D., & Motowildo, S.J. (2001). Personality predictors of citizenship performance. International Journal of Selection and Assessment, 9, 52 69. Boyle, S., Fullerton, J., & Yapp, M. (1993). The rise of the assessment centre: A survey of assessment center usage in the UK. Selection and Development Review, 9(3), 1 3. Bryman, A. (1996). Leadership in organizations. In S.R. Clegg, C. Hardy, & W.R. Nord (Eds.), Handbook of organizational studies (pp. 276292). London: Sage. Burns, J.M. (1978). Leadership. New York: Harper & Row. Campbell, E., Lockyer, C., & Scholarios, D. (1997). Selection practices in Scottish private sector companies: Linkages to organisation size and industrial sector
International Association for Applied Psychology, 2004.
256
BARTRAM
(Occasional Paper 10). Department of Human Resource Management, University of Strathclyde. Campbell, J.P., McHenry, J.J., & Wise, L.L. (1990). Modeling job performance in a population of jobs. Personnel Psychology, 43, 313333. Conway, J.M. (1999). Distinguishing contextual performance from task performance for managerial jobs. Journal of Applied Psychology, 84, 313. Costa, P.T., & McCrae, R.R. (1992). The NEO PI-R Professional Manual. Odessa, FL: Psychological Assessment Resources Inc. Coyne, I., & Bartram, D. (2000). Personnel managers perceptions of dishonesty in the workplace. Human Resource Management Journal, 10(3), 3845. Digman, J.M. (1990). Personality structure: Emergence of the ve-factor model. Annual Review of Psychology, 41, 417 440. Digman, J.M. (1997). Higher-order factors of the big ve. Journal of Personality and Social Psychology, 73, 1246 1256. DiMilia, L., Smith, P.A., & Brown, D.F. (1994). Management selection in Australia: A comparison with British and French ndings. International Journal of Selection and Assessment, 2, 80 90. Donovan, M.A., Drasgow, F., & Probst, T.M. (2000). Does computerizing paperand-pencil job attitude scales make a difference? New IRT analyses offer insight. Journal of Applied Psychology, 85, 305 313. Engelhart, D. (1996). The usage of psychometric tests in France. In M. Smith & V. Sutherland (Eds.), International review of professional issues in selection and assessment ( Vol. 1, pp. 163 164). New York: Wiley. Fiedler, F.E. (1967). A theory of leadership effectiveness. New York: McGraw-Hill. Flanagan, J.C. (1951). Dening the requirements of the executives job. Personnel, 28, 28 35. Funke, U. (1996). German studies in selection and assessment. In M. Smith & V. Sutherland (Eds.), International review of professional issues in selection and assessment (Vol. 2, pp. 169 175). New York: Wiley. Gowing, M.K., & Slivinski, L.W. (1994). A review of North American selection procedures: Canada and the United States of America. International Journal of Selection and Assessment, 2, 103 114. Henley, S., & Bawtree, S. (1996). Training standards and procedures for training psychologists involved in selection and assessment in the United Kingdom. In M. Smith & V. Sutherland (Eds.), International review of professional issues in selection and assessment (Vol. 1, pp. 51 58). New York: Wiley. Hermelin, E., & Robertson, I.T. (2001). A critique and standardization of metaanalytic validity coefcients in personnel selection. Journal of Occupational and Organizational Psychology, 74, 253 278. Hofstede, G. (1991). Cultures and organizations: Software of the mind. London: McGraw-Hill. Hough, L.M., & Oswald, F.L. (2000). Personnel selection: Looking towards the futureremembering the past. Annual Review of Psychology, 51, 631664. House, R.J., Hanges, P.J., Ruiz-Quintanilla, S.A., Dorfman, P.W., Javidan, M., Dickson, M., Gupta, V., & GLOBE (1999). Cultural inuences on leadership and organisations: Project Globe. Advances in Global Leadership, 1, 171 233.
International Association for Applied Psychology, 2004.
ASSESSMENT IN ORGANISATIONS
257
House, R.J., Javidan, M., & Dorfman, P. (2001). Project GLOBE: An introduction. Applied Psychology: An International Review, 50, 479488. Hunt, S.T. (1996). Generic work behaviour: An investigation into the dimensions of entry-level hourly job performance. Personnel Psychology, 49, 5183. Hunter, J.E., & Schmidt, F.L. (1990). Methods of meta-analysis. Newbury Park, CA: Sage. International Test Commission (2001). International guidelines on test use. International Journal of Testing, 2, 93 114. IRS (1997). The state of selection: An IRS survey. Employee Development Bulletin, 85. London: Industrial Relations Services. Irvine, S., & Kyllonen, P. (Eds.) (2002). Item generation for test development. Hillsdale, NJ: Lawrence Erlbaum. Jenkins, A. (2001). Companies use of psychometric testing and the changing demand for skills: A review of the literature. London: Centre for the Economics of Education, London School of Economics and Political Science. Kurz, R., & Bartram, D. (2002). Competency and individual performance: Modelling the world of work. In I.T. Robertson, M. Callinan, & D. Bartram (Eds.), Organisational effectiveness: The role of psychology (pp. 227258). Chichester, UK: Wiley. Levy-Leboyer, C. (1994). Selection and assessment in Europe. In H.C. Triandis, M.D. Dunnette, & L.M. Hough (Eds.), Handbook of industrial and organizational psychology (Vol. 4, pp. 173190). Palo Alto, CA: Consulting Psychologists Press. Lievens, F., & Harris, M.M. (2003). Research on internet recruitment and testing: Current status and future directions. In C.L. Cooper, & I.T. Robertson (Eds.), International Review of Industrial and Organizational Psychology (Vol. 18, pp. 131165). Chichester: John Wiley. Matthews, G. (1997). The Big Five as a framework for personality assessment. In N. Anderson & P. Herriot (Eds.), International handbook of selection and assessment (pp. 475 492). Chichester: Wiley. Murphy, K. (2000). Impact of assessments of validity generalization and situational specicity on the science and practice of personnel selection. International Journal of Selection and Assessment, 8, 194 206. Murphy, K., & Bartram, D. (2002). Recruitment, personnel selection and organisational effectiveness. In I.T. Robertson, M. Callinan, & D. Bartram (Eds.), Organisational effectiveness: The role of psychology (pp. 85114). Chichester, UK: Wiley. Neuman, G., & Baydoun, R. (1998). Computerization of paper-and-pencil tests: When are they equivalent? Applied Psychological Measurement, 22, 7183. Norman, W.T. (1963). Towards an adequate taxonomy of personality attributes: Replicated factor structure in peer nomination personality rating. Journal of Abnormal and Social Psychology, 66, 574 583. OGorman, J.G. (1996). Selection and assessment in Australia. In M. Smith & V. Sutherland (Eds.), International review of professional issues in selection and assessment ( Vol. 1, pp. 25 35). New York: Wiley. Ones, D.S., & Viswesveran, C. (1998). Gender, age and race differences in overt integrity tests: Results across four large-scale job applicant data sets. Journal of Applied Psychology, 83, 35 42.
International Association for Applied Psychology, 2004.
258
BARTRAM
Ones, D.S., & Viswesveran, C. (2001). Integrity tests and other criterion-focused occupational personality scales (COPS) used in personnel selection. International Journal of Selection and Assessment, 9, 3139. Ones, D.S., Viswesveran, C., & Schmidt, F.L. (1993). Comprehensive meta-analysis of integrity test validities: Findings and implications for personnel selection and theories of job performance. Journal of Applied Psychology, 78, 679703. Richman-Hirsch, W.L., Olson-Buchanan, J.B., & Drasgow, F. (2000). Examining the impact of administration medium on examinee perceptions and attitudes. Journal of Applied Psychology, 85, 880 887. Robertson, I.T., Bartram, D., & Callinan, M. (2002). Personnel selection and assessment. In P. Warr (Ed.), Psychology at work (Chapter 5, pp. 100152). London: Penguin Books. Robertson, I.T., & Callinan, M. (1998). Personality and work behaviour. European Journal of Work and Organizational Psychology, 7, 321340. Robertson, I.T., & Kinder, A. (1993). Personality and job competencies: The criterionrelated validity of some personality variables. Journal of Occupational and Organisational Psychology, 66, 225 244. Robertson, I.T., & Smith, M. (2001). Personnel selection. Journal of Occupational and Organizational Psychology, 74, 441 472. Roth, P.L., Bevier, C.A., Bobko, P., Switzer III, F.S., & Tyler, P. (2001). Ethnic group differences in cognitive ability in employment and educational settings: A meta-analysis. Personnel Psychology, 54, 297330. Rowe, P.M., Wiliams, M.C., & Day, A.L. (1994). Selection procedures in North America. International Journal of Selection and Assessment, 2, 7479. Ryan, A.M., McFarland, L., Baron, H., & Page, R. (1999). An international look at selection practices: Nation and culture as explanations of variability in practice. Personnel Psychology, 52, 359 392. Salgado, J.F. (1998). Big Five personality dimensions and job performance in army and civil occupations: A European perspective. Human Performance, 11, 271 288. Salgado, J.F. (1999). Personnel selection methods. In C.L. Cooper & I.T. Robertson (Eds.), International review of industrial and organizational psychology (Vol. 14, pp. 1 54). Chichester, UK: Wiley. Salgado, J.F., Anderson, N., Moscoso, S., Bertua, C., de Fruyt, F., & Rolland, J.P. (2003). A meta-analytic study of General Mental Ability validity for different occupations in the European Community. Journal of Applied Psychology, 88(6), 1068 1081. Saville, P., Sik, G., Nyeld, G., Hackston, J., & MacIver, R. (1996). A demonstration of the validity of the Occupational Personality Questionnaire (OPQ) in the measurement of job competencies across time and in separate organisations. Applied Psychology: An International Review, 45, 243262. Schmidt, F.L., & Hunter, J.E. (1998). The validity and utility of selection methods in personnel psychology: Practical and theoretical implications of 85 years of research ndings. Psychological Bulletin, 124, 262274. Schneider, B. (1996). Whither goest personality at work? Applied Psychology: An International Review, 45, 289 296. Scholarios, D., & Lockyer, C. (1999). Recruiting and selecting professionals: Context, qualities and methods. International Journal of Selection and Assessment, 7, 142156.
International Association for Applied Psychology, 2004.
ASSESSMENT IN ORGANISATIONS
259
Schwartz, S.H. (1999). A theory of cultural values and some implications for work. Applied Psychology: An International Review, 48, 2348. Schwartz, S.H., & Bilsky, W. (1987). Towards a universal psychological structure of human values. Journal of Personality and Social Psychology, 53, 550562. Shackleton, V., & Newell, S. (1991). Management selection: A comparative survey of methods used in top British and French companies. Journal of Occupational Psychology, 64, 23 36. Shackleton, V., & Newell, S. (1994). European management selection methods: A comparison of ve countries. International Journal of Selection and Assessment, 2, 91 102. Shippmann, J.S., Ash, R.A., Battista, M., Carr, L., Eyde, L.D., Hesketh, B., Kehoe, J., Pearlman, K., Prien, E.P., & Sanchez, J.I. (2000). The practice of competency modeling. Personnel Psychology, 53, 703 740. SHL (1992). Motivation questionnaire manual & users guide. Thames Ditton, UK: SHL Group plc. SHL (1993). Inventory of management competencies: Manual and users guide. Thames Ditton, UK: SHL Group plc. SHL (1999). OPQ32: Manual and users guide. Thames Ditton, UK: SHL Group plc. Smith, M., & George, D. (1992). Selection methods. In C.L. Cooper & I.T. Robertson (Eds.), International review of industrial and organizational psychology (Vol. 7, pp. 55 97 ). Chichester, UK: Wiley. Smith, P.B., Dugan, S., & Trompenaars, F. (1996). National culture and the values of organizational employees: A dimensional analysis across 43 nations. Journal of Cross-Cultural Psychology, 27, 231 264. Stanton, J.M. (1999). Validity and related issues in web-based hiring. The Industrial Psychologist (TIP ), 36(3), 69 77. Tett, R.P., Guterman, H.A., Bleier, A., & Murphy, P.J. (2000). Development and content validation of a hyperdimensional taxonomy of managerial competence. Human Performance, 13, 205 251. Vroom, V.H., & Yetton, P.N. (1973). Leadership and decision making. Pittsburgh, PA: University of Pittsburgh Press.