Stubborn Reliance on Intuition and Subjectivity in Employee Selection

Page created by Jorge Hernandez
 
CONTINUE READING
Industrial and Organizational Psychology, 1 (2008), 333–342.
Copyright ª 2008 Society for Industrial and Organizational Psychology. 1754-9426/08

FOCAL ARTICLE

Stubborn Reliance on Intuition and
Subjectivity in Employee Selection

SCOTT HIGHHOUSE
Bowling Green State University

Abstract
The focus of this article is on implicit beliefs that inhibit adoption of selection decision aids (e.g., paper-and-pencil
tests, structured interviews, mechanical combination of predictors). Understanding these beliefs is just as impor-
tant as understanding organizational constraints to the adoption of selection technologies and may be more useful
for informing the design of successful interventions. One of these is the implicit belief that it is theoretically
possible to achieve near-perfect precision in predicting performance on the job. That is, people have an inherent
resistance to analytical approaches to selection because they fail to view selection as probabilistic and subject to
error. Another is the implicit belief that prediction of human behavior is improved through experience. This myth
of expertise results in an overreliance on intuition and a reluctance to undermine one’s own credibility by using
a selection decision aid.

Perhaps the greatest technological achieve-                   unstructured interviews. For example, the
ment in industrial and organizational (I–O)                   right side of Figure 1 shows the results of
psychology over the past 100 years is the                     a meta-analysis conducted on the actual
development of decision aids (e.g., paper-                    effectiveness of these same procedures for
and-pencil tests, structured interviews,                      predicting performance in sales (Vinchur
mechanical combination of predictors) that                    Schippmann, Switzer, & Roth, 1998). Use of
substantially reduce error in the predic-                     any one of the paper-and-pencil tests alone
tion of employee performance (Schmidt &                       outperforms the unstructured interview—a
Hunter, 1998). Arguably, the greatest failure                 procedure that is presumed to assess ability,
of I–O psychology has been the inability to                   personality, and aptitude concurrently.
convince employers to use them. A little over                    Although one might argue that these data
10 years ago, Terpstra (1996) sampled 201                     merely reflect a lack of knowledge about
human resources (HR) executives about the                     effective practice, there is considerable evi-
perceived effectiveness of various selection                  dence that employers simply do not believe
methods. As the left side of Figure 1 shows,                  that the research is relevant to their own sit-
they considered the traditional unstructured                  uation (Colbert, Rynes, & Brown, 2005;
interview more effective than any of the                      Johns, 1993; Muchinsky, 2004; Terpstra &
paper-and-pencil assessment procedures.                       Rozelle, 1997; Whyte & Latham, 1997).
Inspection of actual effectiveness of these                   For example, Rynes, Colbert, and Brown
procedures, however, shows that paper-                        (2002) found that HR professionals were
and-pencil tests commonly outperform                          well aware of the limitations of the unstruc-
                                                              tured interview. Similarly, one of my stu-
                                                              dents conducted a yet-unpublished survey
Correspondence concerning this article should be              of HR professionals (n ¼ 206) about their
addressed to Scott Highhouse. E-mail: shighho@bgsu.edu
   Address: Bowling Green State University, Bowling           views of selection practice. His data indi-
Green, OH 43403.                                              cated that the HR professionals agreed, by

                                                         333
334                                                                                                   S. Highhouse

       Unstructured                                     Unstructured
        Interview                                        Interview

         Specific                                         Specific
       AptitudeTest                                     AptitudeTest

       Personality                                       Personality
          Test                                              Test

        GMA Test                                          GMA Test

                      1      2       3      4       5                  0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
                          Perceived Effectiveness                         Actual Effectiveness (Sales)

Figure 1. Perceived versus actual usefulness of various predictors.
Note. Perceived effectiveness numbers are on a 1–5 scale (1 ¼ not good; 3 ¼ average; 5 ¼
extremely good). Actual effectiveness numbers are correlations corrected for unreliability in
the criterion and range restriction. Because Vinchur, Schippmann, Switzer, and Roth (1998)
did not include interviews, the interview estimate is from Huffcutt and Arthur (1994) level
1 interview. GMA ¼ general mental ability; personality ¼ potency; specific aptitude ¼
sales ability.

a factor of more than 3 to 1, that using tests          Malcolm Gladwell’s (2005) Blink: The
was an effective way to evaluate a candi-               Power of Thinking Without Thinking and
date’s suitability and that tests that assess           Gerd Gigerenzer’s (2007) Gut Feelings:
specific traits are effective for hiring em-            The Intelligence of the Unconscious, which
ployees. At the same time, however, these               extol the virtues of intuitive decision mak-
same professionals agreed, by more than 3               ing. Although the assertions of these authors
to 1, that you can learn more from an infor-            have little relevance for the prediction of
mal discussion with job candidates and that             human performance, the popularity of their
you can ‘‘read between the lines’’ to detect            work likely reinforces the common belief
whether someone is suitable to hire. This               that good hiring is a matter of experience
apparent conflict between knowledge and                 and intuition.
belief seems loosely analogous to the com-
mon practice of preferring brand name cold
                                                        Implicit Beliefs
remedies to store brand remedies containing
the same ingredients. People know that the              My colleagues and I (Lievens, Highhouse, &
store brands are identical, but they do not             De Corte, 2005) conducted a policy-capturing
trust them for their own colds.                         study of the decision processes of retail man-
   Some might argue that the tide is turning.           agers making hypothetical hiring decisions.
Much has been written on the merits of evi-             We found that the managers placed more
dence-based management (Pfeffer & Sutton,               emphasis on competencies assessed by
2006; Rousseau, 2006). This approach,                   unstructured interviews than on competen-
much like evidence-based medicine, relies               cies measured by tests, regardless of what
on the best available scientific evidence to            those competencies were. They placed more
make decisions. At the core of this move-               emphasis, for instance, on Extraversion than
ment is ‘‘analytics’’ or data-based decision            on general mental ability when Extraversion
making (e.g., Ayers, 2007). Discussions of              was assessed using an unstructured inter-
number crunching in the arena of personnel              view (and general mental ability was as-
selection, however, are almost always lim-              sessed using a paper-and-pencil test). The
ited to anecdotes from professional sports              opposite was found when Extraversion was
(e.g., Davenport, 2006). Competing with                 assessed using a paper-and-pencil test and
the analytical point of view are books like             general mental ability was assessed using
Reliance on intuition and subjectivity                                                                      335

an unstructured interview! Clearly, these             correct the imperfections (rather than exac-
managers believed that good old-fashioned             erbate them). The court’s majority opinion
‘‘horse sense’’ was needed to accurately size         in Gratz suggests that individualized meth-
up applicants (see Phelan & Smith, 1958).             ods of selection are more fair and reliable
   The reluctance of employers to use ana-            than impersonal ‘‘mechanical’’ ones. Both
lytical selection procedures is at least              of these examples illustrate two implicit
partially a reflection of broader misconcep-          beliefs about employee selection: (1) people
tions that the general public has about how           believe that it is possible to achieve near-
to go about assessing and selecting people            perfect precision in the prediction of
for jobs. Consider two high-profile policy            employee success, and (2) people believe
opinions on testing and selection in the              that there is such a thing as intuitive expertise
United States.                                        in the prediction of human behavior. These
                                                      implicit beliefs exert their influence on pol-
   d   In 1990, the National Commission on            icy and practice, even though they may not
       Testing and Public Policy (1990) issued        be immediately accessible (Kahneman,
       eight recommendations for testing in           2003). I acknowledge that there are a num-
       schools and the workplace. Among               ber of contextual reasons for resistance to
       those was the statement as follows:            selection technologies, including organiza-
       ‘‘Test scores are imperfect measures           tional politics, habit, and culture, along with
       and should not be used alone to make           the existing legal climate (e.g., Johns, 1993;
       important decisions about individuals’’        Muchinsky, 2004). However, whereas con-
       (National Commission on Testing and            textual issues are often situation specific,
       Public Policy, 1990, p. 30). The com-          these are universal ‘‘truths’’ about people.
       mission’s chairman, Bernard Gifford of         As such, understanding and studying them
       Apple Computer, commented, ‘‘We                provides hope for overcoming user resis-
       just believe that under no circumstan-         tance to selection decision aids.
       ces should individuals be denied a job
       or college admission exclusively based
                                                      Irreducible Unpredictability
       on test scores’’ (‘‘Panel Criticizes Stan-
       dard Testing,’’ 1990).                         I recently came across an article in a popular
   d   In the landmark Supreme Court deci-            trade magazine for executives, purportedly
       sion on affirmative action at the Univer-      summarizing the state of the science on
       sity of Michigan, Justice Rehnquist            executive assessment (Sindelar, 2002). I
       concluded that consideration of race           was struck by a statement made by the
       as a factor in student admission is            author: ‘‘For many top-level positions, tech-
       acceptable—but it must be done at              nical competence accounts for only 20
       the individual level, with each appli-         percent of a successful alignment. Psycho-
       cant considered holistically. In concur-       logical factors account for the rest’’ (pp.
       rence, Justice O’Connor commented,             13–14).1 Whether intentional or not, the
       ‘‘But the current [student selection] sys-     author was clearly implying what is shown
       tem, as I understand it, is a nonindivi-       on the top of Figure 2—that 80% of the var-
       dualized, mechanical one. As a result, I       iance in executive success can be explained
       join the Court’s opinion . . . .’’ (Gratz v.   by psychological factors (presumably tem-
       Bollinger, 2003, Concurrence 1).               perament or personality). Reality, however,
                                                      is much more like the chart on the bottom
Although these positions sound reasonable             of Figure 2—showing that most of the vari-
on the surface, they represent fundamen-              ance in executive success is simply not
tally flawed assumptions. No one disputes
that test scores are imperfect measures,
but the testing commission implies that               1. The author identified his affiliation as the ‘‘Institute
combining them with something else will                  for Advanced Business Psychology.’’
336                                                                                       S. Highhouse

                                Tech.                success. Campbell wrote: ‘‘No external
                              Competence
                                 20%
                                                     source imposed this [validity ceiling] stan-
                                                     dard on the discipline or even argued that
                                                     there should be a standard at all’’ (p. 689).
                                                         Recall the earlier comment by the
                                                     national testing commission, cautioning
                                                     that tests are ‘‘imperfect’’ and must be
                                                     supplemented with other things. It is remark-
Psych. Factors
                                                     ably similar to Viteles’ (1925) observation
    80%                                              that ‘‘objective scores of vocational tests
                                                     are at best uncertain diagnostic criteria’’
                               Tech.
                             Competence              (p. 132). This early pioneer of I–O was argu-
                                20%                  ing that standardized methods of assessment
                                                     could only fill the proverbial glass halfway.
                                                     Intuitive judgment was needed to fill it the
                                    Psych. Factors   rest of the way. Viteles wrote: ‘‘It is the opin-
                                        10%
                                                     ion of the writer that in the cause of greater
                                                     scientific accuracy in vocational selection
Unpredictability
    70%                                              in industry the statistical point of view must
                                                     be supplemented by a clinical point of view’’
                                                     (p. 134). Countering this position was Freyd
Figure 2. Variance in success accounted for
                                                     (1926), who cautioned against allowing intu-
by technical competence and psychological
                                                     ition to creep into hiring decisions. Freyd,
factors.
                                                     who represented the analytical viewpoint
                                                     of selection, argued ‘‘allowing selection to
predictable prior to employment. The busi-           be influenced by personal interpretations
ness of assessment and selection involves            with their unavoidable prejudices instead of
considerable irreducible unpredictability;           relying upon objective measures gives even
yet, many seem to believe that all failures          less consideration to the well-being and
in prediction are because of mistakes in the         interest of the individual worker’’ (p. 354).
assessment process. Put another way, people          History proved Freyd prescient.
seem to believe that, as long as the applicant           Table 1 shows the results of the earliest
is the right person for the job and the ap-          study investigating the relative effectiveness
plicant is accurately assessed, success is           of standardized procedures alone versus
certain. The ‘‘validity ceiling’’ has been a con-    supplementing those procedures with intu-
tinually vexing problem for I–O psychology           itive judgment (Sarbin, 1943). As you can
(see Campbell, 1990; Rundquist, 1969). Enor-         see, academic achievement was better pre-
mous resources and effort are focused on the         dicted by the standardized scores alone
quixotic quest for new and better predictors         than by the scores plus clinical judgment.
that will explain more and more variance in          The notion that analysis outperforms intui-
performance. This represents a refusal, by           tion in the prediction of human behavior is
knowledgeable people, to recognize that              among the most well-established findings
many determinants of performance are not             in the behavioral sciences (Grove & Meehl,
knowable at the time of hire. The notion that        1994; Grove, Zald, Lebow, Snitz, & Nelson,
it is still possible to achieve large gains in the   2000).2 Why, therefore, does the intuitive
prediction of employee success reflects a fail-
ure to accept that there is no such thing as
perfect prediction in this domain. Campbell          2. Although few studies in I–O have explicitly made
noted that our poor professional self-esteem            this comparison, there are a number of examples
                                                        where tests alone outpredicted tests 1 intuition
is based on an unrealistic notion of what can           (e.g., Borneman, Cooper, Klieger, & Kuncel, 2007;
be achieved in the prediction of employee               Huse, 1962; Meyer, 1956).
Reliance on intuition and subjectivity                                                           337

Table 1. Sarbin’s (1943) Investigation of Two Methods for Predicting Success of University
of Minnesota Undergraduates Admitted in 1939
Predictor composite                                                     Correlation with criterion (r)
High school rank 1 college aptitude test                                             .45
High school rank 1 college aptitude test 1                                           .35
  intuitive judgment of counselors

perspective remain so appealing? Einhorn             (1986) noted, however, that one must be
(1986) observed that a crucial distinction           willing to accept error to make less error.
between the intuitive and the analytical
approaches to human prediction is the
                                                     Myth of Expertise
worldview of the people making the judg-
ments. According to Einhorn, the intuitive           I have argued that one of the reasons that
approach reflects a deterministic world-             people have an inherent resistance to analyt-
view, one that rejects the idea that the future      ical approaches to hiring is that they fail to
is inherently probabilistic. This is con-            view selection in probabilistic terms. A
trasted with the analytical worldview,               related but different reason for employer ret-
which accepts uncertainty as inevitable.             icence to use selection decision aids is that
Consider the San Diego Chargers profes-              most people believe in the myth of selection
sional football team who, despite having             expertise. By this I mean the belief that one
a regular season record of 14-2 in 2006,             can become skilled in making intuitive judg-
fired its head coach following a play-off            ments about a candidate’s likelihood of suc-
loss. The fired coach had a reputation for           cess. This is reflected in the survey responses
leading teams to successful regular season           of the HR professionals who believed in
records, only to lose the big games. The             ‘‘reading between the lines’’ to size up job
Chargers organization evidently failed to            candidates. It is also evidenced in the pheno-
consider that the contribution of uncer-             menal growth of the professional recruiter
tainty to a play-off outcome is much greater         or ‘‘headhunter’’ profession (Finlay & Cover-
than to a 16-game season record. Abelson             dill, 1999) and the perseverance of the
(1985) found that knowledgeable baseball             holistic approach to managerial assessment
fans overestimated by a factor of 75 the con-        (Highhouse, 2002).
tribution of skill (vs. chance) to the likeli-           Despite this widespread belief in intuitive
hood of a major league baseball player               expertise, the data suggest that it is a myth.
getting a hit in a given turn at bat.                For example, the considerable research on
    Intuitive approaches to employee selec-          predicting human behavior per se shows that
tion make the errors in selection ambiguous.         experience does not improve predictions
Analytical approaches make them part of              made by clinicians, social workers, parole
the process—hence, visible. Considerable             boards, judges, auditors, admission com-
research suggests that ambiguity about the           mittees, marketers, and business planners
likelihood of an outcome (e.g., the operation        (Camerer & Johnson, 1991; Dawes, Faust,
has an unknown chance of success) encour-            & Meehl, 1989; Grove et al., 2000; Sherden,
ages more optimism than a low known prob-            1998). Although it is commonly accepted
ability (e.g., the operation has a 20% chance        that some (employment) interviewers are
of success; see Kuhn, 1997). There is little         better than others, research on variance in
room for optimism when a composite of pre-           interviewer validity suggests that differences
dictors is known to leave 75% of the variance        are due entirely to sampling error (Pulakos,
unexplained. This may explain why selection          Schmitt, Whitney, & Smith, 1996). Exist-
procedures that are difficult to evaluate (e.g.,     ing evidence suggests that the interrater
feelings about ‘‘fit’’) are so attractive. Einhorn   reliability of the traditional (unstructured)
338                                                                                           S. Highhouse

interview is so low that, even with a perfectly             expert’s ability to interpret configurations
reliable and valid criterion, interview-based               of traits (Prien, Schippmann, & Prien, 2003).
judgments could never account for more                      The notion behind this argument is that
than 10% of the variance in job performance                 each candidate is unique, and one must con-
(Conway, Jako, & Goodman, 1995).3 This                      sider each piece of information about the
empirical evidence is troubling for a proce-                candidate in light of all the other pieces of
dure that is supposed to simultaneously take                information. In other words, assessing pat-
into account ability, motivation, and person–               terns of traits is more accurate than assessing
organization fit. Keep in mind also that these              traits individually. For example, Prien et al.
findings are based on interviews that had rat-              noted that executive assessment requires a
ings associated with the interviewers’ judg-                ‘‘dynamic interpretation’’ of applicant data,
ments. Thus, the unstructured interviews                    one that takes into account interactions
subjected to meta-analyses are almost cer-                  between test scores and other observations
tainly unusual and on the high end of rigor.                (p. 125). This view is reinforced by leadership
The data do not paint a sanguine picture of                 theorists who assert that leader characteristics
intuitive judgment in the hiring process.                   exhibit complex configural relations with
   There are commonly two scholarly rebut-                  leadership outcomes (e.g., Zaccaro, 2007).
tals to the arguments against prediction                        Even if we do accept that decision makers
expertise. I will consider these in turn. One               incorporate broken-leg cues and configura-
response to the limitations of intuitive                    tions of traits, existing evidence suggests that
approaches to selection is to focus on the                  these things account for negligible variance
ability of experts to spot idiosyncrasies in                in the predicted outcome. For example,
a candidate’s profile (Jeanneret & Silzer,                  Dawes (1971) modeled admission decisions
1998). Meehl (1954) noted that one limita-                  of a four-person graduate admissions com-
tion of analytical formulas was their inability             mittee using a bootstrapping procedure. This
to incorporate ‘‘broken-leg’’ cues. The term                is shown in Figure 3. Dawes found that the
comes from an anecdotal example in which                    model (i.e., paramorphic representation) of
one is trying to predict whether or not a per-              the admission committee’s judgments out-
son will go to the movie on a particular day.               performed the committee itself. More rele-
A mechanical formula might take into                        vant to this discussion, however, was the fact
account things like the nature of the movie                 that, whereas a linear combination of the
(e.g., less likely to go to romantic comedy) or             expert cues correlated significantly (r ¼ .25)
the weather (e.g., more likely to go on a rainy             with the criterion, the residual—which in-
day). The mechanical procedure would not                    cluded configural judgments, broken-leg
take into account, however, an event that is                cues, and error—was inconsequential (r ¼
extremely rare (e.g., the person has a broken               .01). Camerer and Johnson (1991) noted
leg), and thus, the mechanical prediction                   that, despite accounting for a large portion
will not be as accurate as a prediction based               of the error term, broken-leg cues and con-
on a simple intuitive observation. A mechan-                figural judgments consistently provide little
ical approach to selection would not, the                   incremental gain in prediction—even for so-
logic goes, consider idiosyncratic charac-                  called experts. The problem with broken-leg
teristics of any particular job candidate—a                 cues is that people rely too much on them
seasoned expert would.                                      because they present compelling stories.
   Another common response to criticisms                    The tendency to be seduced by detailed
of intuitive selection is to focus on the                   stories causes people to ignore relevant
                                                            information and to violate simple rules of
                                                            logic (see Highhouse, 1997, 2001). Also, as
3. Meta-analysis suggests that it accounts for negligible   one reviewer noted, broken legs are them-
   incremental validity over simple paper-and-pencil        selves constructs that can and should be
   tests of cognitive ability and conscientiousness
   (Cortina, Goldstein, Payne, Davison, & Gilliland,        measured reliably. The problem with trait
   2000).                                                   configurations, on the other hand, is that
Reliance on intuition and subjectivity                                                                  339

                                   Bootstrapped Models of Experts

                                               r = .19
                               Predicted                       Expert
                               Outcome                       Predictions

                                   r = .25
                                              Model of
                                                              linear combination of cues
                   r = .01                     Expert

                                                              configural judgments

                                             Residuals        “broken-leg” cues

                                                              error

Figure 3. Results from Dawes’ (1971) examination of graduate admissions decisions.

they require feats of information integration        mentators hold the Bowl Championship
that contradict current understanding of             Series, which is a mechanical formula that
human cognitive limitations (Ruscio,                 incorporates expert ratings (e.g., coaches
2003). And true real-world examples of pre-          poll) and computer rankings (e.g., wins and
dictive interactions between job applicant           losses of opponents) into an overall ranking
characteristics are difficult to find (e.g.,         of football teams. The nature of the com-
Sackett, Gruys, & Ellingson, 1998). Hastie           plaints (‘‘unplug the computers’’) suggests
and Dawes (2001) distilled from the vast lit-        that people do not want mechanical formu-
erature on prediction ‘‘experts’’ the following      las making their expert decisions about who
stylized facts:                                      attends bowl games. A University of Oregon
                                                     coach infamously declared: ‘‘I liken the BCS
   d   They rely on few pieces of information.       to a bad disease, like cancer’’ (Vondersmith,
   d   They lack insight into how they arrive at     2001). Another example of this bias against
       predictions.                                  decision aids is the considerable patient
   d   They exhibit poor interjudge agree-           resistance to diagnostic decision aids (Arkes,
       ment.                                         Shaffer, & Medow, 2007). Arkes and his col-
   d   They become more confident in their           leagues found that physicians who made
       accuracy when irrelevant information          computer-based diagnoses of ankle injuries
       is presented.                                 were perceived less competent, profes-
                                                     sional, and thorough than physicians who
The obvious remedy to the limitations of             made diagnoses without any aids. Indeed,
expertise is to structure expert intuition and       the idea that (with the appropriate data)
mechanically combine it with other decision          a physician might not even need to meet or
aids, such as paper-and-pencil inventories.          interact with a patient to understand his or
However, there would likely be consider-             her personal health issues would be a hard
able resistance to structuring or mechaniz-          sell to most people. Physicians, aware of this
ing the judgment process (e.g., Lievens et al.,      lay bias against ‘‘cookbook medicine,’’
2005; van der Zee, Bakker, & Bakker, 2002).          grossly underutilize these valuable technol-
Most people believe that aspects of an appli-        ogies in practice (Kaplan, 2001).4 Hastie
cant’s character are far too complex to be           and Dawes (2001) noted that relying on
assessed by scores, ratings, and formulas.
   An example of the irrationality of this bias
                                                     4. This underutilization also results from overconfi-
against decision aids is the contempt with              dence on the part of physicians in their own diagnos-
which most college football fans and com-               tic expertise.
340                                                                                        S. Highhouse

expertise is more socially acceptable than            zation. It was at this point a senior com-
relying on test scores or formulas. Research          pany official said to me, ‘‘I fail to see the
on medical decision making supports this              basis for your enthusiasm.’’ (p. 194)
contention. It is no wonder, therefore, that
HR practitioners would be reluctant to                Research on probability neglect (Sunstein,
undermine their status by administering            2002) suggests that people make little dis-
a paper-and-pencil test, structuring an            tinction between probabilities that they
employment interview, or plugging ratings          consider small. In addition, research on
into a mechanical formula.                         evaluability (Hsee, 1996) has shown that
                                                   most attributes cannot be evaluated without
                                                   appropriate context. Perhaps if Muchinsky
Concluding Remarks
                                                   (2004) had compared his .50 to flipping
We know quite a bit about applicant reac-          a coin (.00) or to an unstructured interview
tions to hiring methods (Hausknecht, Day, &        (.20), management would have been more
Thomas, 2004), but very little attention has       impressed. Perhaps management would
been given to user resistance to selection         have been more impressed by a common-
decision aids. Campbell (1990) noted: ‘‘We         language effect size indicator or by an
still do not know much about how to best           expectancy chart. We simply do not have
communicate selection results to people            the research to guide these communication
outside the [I-O] profession’’ (p. 704). Fifteen   decisions.
years later, Anderson (2005) lamented: ‘‘In           The traditional unstructured interview has
fact, the whole area of practitioner beliefs       remained the most popular and widely used
about selection methods and processes is           selection procedure for over 100 years
a gargantuan one which research has made           (Buckley, Norris, & Wiese, 2000). This is
little or no inroads into’’ (p. 19). I have        despite the fact that, during this same period,
inferred from the general psychological lit-       there have been significant advancements in
erature, and the specific selection literature,    the development of selection decision aids.
two implicit beliefs that likely inhibit the       Guion (1965) argued that the waste of
widespread acceptance of selection tech-           human resources caused by poor selection
nologies. These include the belief that it is      procedures should pain the professional
possible to achieve near-perfect precision in      conscience of I–O psychologists. It is true
predicting performance on the job and the          that people are not very predictable, but
belief that intuitive prediction can be            selection decision aids help.
improved by experience. People trust that
the complex characteristics of applicants
                                                   References
can be best assessed by a sensitive, equally
complex human being. This does not stand           Abelson, R. P. (1985). A variance explanation paradox:
                                                      When a little is a lot. Psychological Bulletin, 97,
up to scientific scrutiny, and I–O psycholo-          128–132.
gists need to begin focusing their efforts on      Anderson, N. (2005). Relationship between practice
understanding how to navigate these waters.           and research in personnel selection: Does the left
                                                      hand know what the right is doing? In A. Evers, N.
We can begin by drawing from the judgment             Anderson, & O. Voskuijl (Eds.), The Blackwell hand-
and decision making and human factors lit-            book of personnel selection (pp. 1–24). Malden,
eratures on how to better communicate                 MA: Blackwell.
                                                   Arkes, H., Shaffer, V. A., & Medow, M. A. (2007).
uncertainty and error. We also need to learn          Patients derogate physicians who use a computer-
how to better calibrate user expectations.            assisted diagnostic aid. Medical Decision Making,
Consider Muchinsky’s (2004) experience in             27, 189–202.
                                                   Ayers, I. (2007). Super crunchers: Why thinking-by-
communicating a .50 validity coefficient for          numbers is the new way to be smart. New York:
a mechanical comprehension test:                      Random House.
                                                   Borneman, M. J., Cooper, S. R., Klieger, D. M., & Kuncel,
                                                      N. R. (2007, April). The efficacy of the admissions
   my pleasure regarding the findings was             interview: A meta-analysis. In N. R. Kuncel (Chair),
   highly apparent to the client organi-              Alternative predictors of academic performance:
Reliance on intuition and subjectivity                                                                               341

    The glass is half empty. Symposium conducted at           Hausknecht, J. P., Day, D. V., & Thomas, S. C. (2004).
    the Annual Meeting of the National Council on                Applicant reactions to selection procedures: An
    Measurement in Education, Chicago, IL.                       updated model and meta-analysis. Personnel Psy-
Buckley, M. R., Norris, A. C., & Wiese, D. S. (2000). A          chology, 57, 639–683.
    brief history of the selection interview: May the next    Highhouse, S. (1997). Understanding and improving
    100 years be more fruitful. Journal of Management            job-finalist choice: The relevance of behavioral
    History, 6, 113–126.                                         decision research. Human Resource Management
Camerer, C. F., & Johnson, E. J. (1991). The process-per-        Review, 7, 449–470.
    formance paradox in expert judgment: How can              Highhouse, S. (2001). Judgment and decision mak-
    experts know so much and predict so badly? In K.             ing research: Relevance to industrial and orga-
    A. Ericsson & J. Smith (Eds.), Toward a general theory       nizational psychology. In N. Anderson, D. S. Ones,
    of expertise: Prospects and limits (pp. 195–217).            H. K. Sinangil, & C. Viswesvaran (Eds.), Handbook
    Cambridge, UK: Cambridge University Press.                   of industrial, work and organizational psychology
Campbell, J. P. (1990). Modeling the performance                 (pp. 314–332). Thousand Oaks, Sage.
    prediction problem in industrial and organizational       Highhouse, S. (2002). Assessing the candidate as
    psychology. In M. D. Dunnette & L. M. Hough (Eds.),          a whole: A historical and critical analysis of individ-
    Handbook of industrial and organizational psychol-           ual psychological assessment for personnel decision
    ogy (Vol. 1, 2nd ed., pp. 687–732). Palo Alto, CA:           making. Personnel Psychology, 55, 363–396.
    Consulting Psychologists Press.                           Hsee, C. K. (1996). The evaluability hypothesis: An
Colbert, A. E., Rynes, S. L., & Brown, K. G. (2005). Who         explanation for preference reversals between joint
    believes us? Understanding managers’ agreement               and separate evaluations of alternatives. Organiza-
    with human resource findings. Journal of Applied             tional Behavior and Human Decision Processes, 67,
    Behavioral Science, 41, 304–325.                             242–257.
Conway, J. M., Jako, R. A., & Goodman, D. F. (1995). A        Huffcutt, A. L., & Arthur, W (1994). Hunter and Hunter
    meta-analysis of interrater and internal consistency         (1984) revisited: Interview validity for entry-level
    reliability of selection interviews. Journal of Applied      jobs.
    Psychology, 80, 565–579.                                  Huse, E. F. (1962). Assessments of higher level personnel
Cortina, J. M., Goldstein, N. B., Payne, S. C., Davison, H.      IV: The validity of assessment techniques based on
    K., & Gilliland, S. W. (2000). The incremental val-          systematically varied information. Personnel Psy-
    idity of interview scores over and above cognitive           chology, 15, 195–205.
    ability and conscientiousness scores. Personnel           Jeanneret, R., & Silzer, R. (1998). An overview of psy-
    Psychology, 53, 325–351.                                     chological assessment. In R. Jeanneret & R. Silzer
Davenport, T. H. (2006, January). Competing on analy-            (Eds.), Individual psychological assessment: Predict-
    tics. Harvard Business Review, 84, 99–107.                   ing behavior in organizational settings (pp. 3–26).
Dawes, R. M. (1971). A case study of graduate admis-             San Francisco: Jossey-Bass.
    sions: Application of three principles of human           Johns, G. (1993). Constraints on the adoption of psy-
    decision making. American Psychologist, 26,                  chology-based personnel practices: Lessons from
    180–188.                                                     organizational innovation. Personnel Psychology,
Dawes, R. M., Faust, D., & Meehl, P. E. (1989). Clinical         46, 569–592.
    versus actuarial judgment. Science, 243, 1668–            Kahneman, D. (2003). A perspective on judgment and
    1774.                                                        choice: Mapping bounded rationality. American
Einhorn, H. J. (1986). Accepting error to make less error.       Psychologist, 58, 697–720.
    Journal of Personality Assessment, 50, 387–395.           Kaplan, B. (2001). Evaluating informatics applications:
Finlay, W., & Coverdill, J. E. (1999). The search game:          Clinical decision support systems literature review.
    Organizational conflicts and the use of headhunters.         International Journal of Medical Informatics, 64, 15–37.
    Sociological Quarterly, 40, 11–30.                        Kuhn, K. (1997). Communicating uncertainty: Framing
Freyd, M. (1926). The statistical viewpoint in vocational        effects on responses to vague probabilities. Organi-
    selection. Journal of Applied Psychology, 10, 349–           zational Behavior and Human Decision Processes,
    356.                                                         71, 55–83.
Gigerzenger, G. (2007). Gut feelings: The intelligence        Lievens, F., Highhouse, S., & De Corte, W. (2005). The
    of the unconscious. London: Penguin Books.                   importance of traits and abilities in supervisors’ hir-
Gladwell, M. (2005). Blink: The power of thinking with-          ability decisions as a function of method of assess-
    out thinking. New York: Little, Brown and Company.           ment. Journal of Occupational and Organizational
Gratz v. Bollinger, 539 U.S. 244 (2003).                         Psychology,78, 453–470.
Grove, W. M., & Meehl, P. E. (1994). Comparative effi-        Meehl, P. E. (1954). Clinical versus statistical prediction:
    ciency of informal (subjective, impressionistic) and         A theoretical analysis and a review of the evidence.
    formal (mechanical, algorithmic) prediction proce-           Minneapolis, MN: University of Minnesota.
    dures: The clinical–statistical controversy. Psychol-     Meyer, H. H. (1956). An evaluation of a supervisory
    ogy, Public Policy, and Law, 2, 293–323.                     selection program. Personnel Psychology, 9, 499–
Grove, W. M., Zald, D. H., Lebow, B. S., Snitz, B. E., &         513.
    Nelson, C. (2000). Clinical versus mechanical pre-        Muchinsky, P. M. (2004). When the psychometrics of
    diction. Psychological Assessment, 12, 19–30.                test development meets organizational realities: A
Guion, R. M. (1965). Industrial psychology as an aca-            conceptual framework for organizational change,
    demic discipline. American Psychologist, 20, 815–            examples, and recommendations. Personnel Psy-
    821.                                                         chology, 57, 175–209.
Hastie, R., & Dawes, R. M. (2001). Rational choice in an      National Commission on Testing and Public Policy.
    uncertain world. Thousand Oaks, CA: Sage.                    (1990). From gatekeeper to gateway: Transforming
342                                                                                                      S. Highhouse

    testing in America. Chestnut Hill, MA: National Com-      Sarbin, T. L. (1943). A contribution to the study of
    mission on Testing and Public Policy, Boston College.         actuarial and individual methods of prediction.
Panel criticizes standard testing. (1990, May 24). The            American Journal of Sociology, 48, 598–602.
    New York Times. Retrieved December 10, 2007,              Schmidt, F. L., & Hunter, J. E. (1998). The validity and
    from query.nytimes.com                                        utility of selection methods in personnel psychol-
Pfeffer, J., & Sutton, R. I. (2006). Hard facts, danger-          ogy: Practical and theoretical implications of 85
    ous half-truths and total nonsense: Profiting from            years of research findings. Psychological Bulletin,
    evidence-based management. Boston: Harvard                    124, 262–274.
    Business School Press.                                    Sherden, W. (1998). The fortune sellers: The big busi-
Phelan, J., & Smith, G. W. (1958). To select execu-               ness of buying and selling predictions. New York:
    tives—combine interviews, tests, horse sense. Per-            John Wiley.
    sonnel Journal, 36, 417–421.                              Sindelar, S. (2002, October). Executive assessment.
Prien, E. P., Schippmann, J. S., & Prien, K. O. (2003).           Executive Excellence, 19, 13–14.
    Individual assessment: As practiced in industry and       Sunstein, C. R. (2002). Probability neglect: Emotions,
    consulting. Mahwah, NJ: Lawrence Erlbaum.                     worst cases, and law. Yale Law Journal, 112, 61–107.
Pulakos, E. D., Schmitt, N., Whitney, D., & Smith, M.         Terpstra, D. E. (1996, May). The search for effective
    (1996). Individual differences in interviewer ratings:        methods. HR Focus, 73, 16–17.
    The impact of standardization, consensus discus-          Terpstra, D. E., & Rozelle, J. (1997). Why some poten-
    sion, and sampling error on the validity of a struc-          tially effective staffing practices are seldom used.
    tured interview. Personnel Psychology, 49, 85–102.            Public Personnel Management, 26, 483–495.
Rousseau, D. M. (2006). Is there such a thing as evi-         van der Zee, K. I., Bakker, A. B., & Bakker, P. (2002). Why
    dence-based management? Academy of Manage-                    are structured interviews so rarely used in personnel
    ment Review, 31, 256–269.                                     selection? Journal of Applied Psychology, 87, 176–184.
Rundquist, E. A. (1969). The prediction ceiling. Person-      Vinchur, A. J., Schippmann, J. S., Switzer, F. S., & Roth, P.
    nel Psychology, 22, 109–116.                                  L. (1998). A meta-analytic review of predictors of job
Ruscio, J. (2003). Holistic judgment in clinical practice:        performance for salespeople. Journal of Applied
    Utility or futility? Scientific Review of Mental Health       Psychology, 83, 586–597.
    Practice, 2, 38–48.                                       Viteles, M. S. (1925). The clinical viewpoint in vocational
Rynes, S. L., Colbert, A. E., & Brown, K. G. (2002). HR           selection. Journal of Applied Psychology, 9, 131–138.
    professionals’ beliefs about effective human              Vondersmith, J. (2001, December 11). Bellotti sounds
    resource practices: Correspondence between                    off, favors playoffs. The Portland Tribune. Retrieved
    research and practice. Human Resources Manage-                January 10, 2008, from www.portlandtribune.com
    ment, 41, 149–174.                                        Whyte, G., & Latham, G. (1997). The futility of utility
Sackett, P. R., Gruys, M. L., & Ellingson, J. E. (1998).          analysis revisited: When even an expert fails. Per-
    Ability-personality interactions when predicting              sonnel Psychology, 50, 601–610.
    job performance. Journal of Applied Psychology,           Zaccaro, S. J. (2007). Trait-based perspectives of leader-
    83, 545–556.                                                  ship. American Psychologist, 62, 6–16.
You can also read