2.3 . 3?}, aw umnmbukm . Jim man? 5: . $1 . 25.... I. I.» . 3.: 5!! a: ..I::);. I! ‘2; 1 a... 7.! ‘1: but...) Mquugufnu-T-u 'w v I . "won-9'!” épfigévfzifi . . , . . gfigsmsifiéfi [II E. E. ‘ . .3125 JN__xmmu THESIS 7 4'- ’-" OC F .g,’ . ‘-_. IHIHJHIHUHJHIHUHIIHIWHIMllJlllllllllllmllll 193 02048 6530 LIBRARY Michigan State University This is to certify that the dissertation entitled TOWARD AN INTEGRATED MODEL OF APPLICANT FAKING presented by Lynn A. McFarland has been accepted towards fulfillment of the requirements for Ph . D . degree in W Date l'7L/ 3‘ 5/00 MS U is an Affirmative Action/Equal Opportunity Institution 0. 12771 PLACE IN RETURN BOX to remove this checkout from your record. To AVOID FINES return on or before date due. MAY BE RECALLED with earlier due date if requested. EAJEDDUE DATE DUE “DATE DUE [In ‘wvv1 DEC 1 9 zbua moo c/CIHCJDflDtIJ55-p.“ TOWARD AN INTEGRATED MODEL OF APPLICANT FAKING By Lynn A. McFarland A DISSERTATION Submitted to Michigan State University In partial fulfillment of the requirements For degree of DOCTOR OF PHILOSOPHY Department of Psychology 2000 .1“: ABSTRACT TOWARD AN INTEGRATED MODEL OF APPLICANT FAKIN G Bv J Lynn A. McFarland The present study used the theory of planned behavior to predict faking on a personality selection test. Participants’ attitudes toward faking, subjective norms toward faking, perceived behavioral control over faking, and intentions to fake a selection test were all assessed. In addition, participants were asked to take the personality measure under instructions to be honest and also under instructions to respond like an applicant. Their faking behavior was assessed through the use of difference scores and a social desirability scale. The theory of planned behavior predicted both the intention to fake the selection test and actual faking behavior. However, proposed moderators of these relationships, such as valence toward performing well on the test, warning of a lie scale, and knowledge of the constructs assessed by the personality test, were generally not supported. Practical implications of these results and directions for future research are discussed. o' i-‘ n'“ ~..L -. C. SI ACKNOWLEDGMENTS I want to everyone who made this dissertation possible, both directly by providing feedback and advice on the dissertation, and indirectly for providing support and encouragement to me personally throughout the process. First, thanks to Ann Marie Ryan, my chair, for providing guidance and support throughout the dissertation process. I also want to thank my committee members, Drs. Neal Schmitt, Rick DeShon, Dan Ilgen, and Murray Barrick for their insightful comments and recommendations. A thank you goes to Melissa Kabin for helping me collect my data and to Christine Scheu for helping me to get those tasks done that were so difficult to do from Maryland. I also thank Christine for her general support and friendship along the way. I want to thank my family, especially my parents, William and Rosanne McFarland. They have always provided me with love and encouragement and have been there for me in the toughest of times cheering me on. Without their constant support I would not have been able to do any of this. I also want to thank my parents for taking the time to learn what industrial/organizational psychology is. This is a true demonstration of their love, concern, and support, and I am sincerely touched. Finally, most of all, I want to thank my husband, Robert Ployhart, for his advice, support, friendship, and love. He read this dissertation almost as many times as I did and I greatly appreciate it! iii TABLE OF CONTENTS LIST OF TABLES .............................................................................................................. v LIST OF FIGURES ............................................................................................................ vi INTRODUCTION ............................................................................................................... 1 Definitions and Why People Fake ............................................................................... 3 Evidence that Applicants Can and Do Fake ................................................................ 5 Influence of Faking on the Results of Personality Measures .................................... 10 A Model of Faking .................................................................................................... 30 Method .......................................................................................................................... 47 Sample ....................................................................................................................... 47 Design ........................................................................................................................ 47 Procedure ................................................................................................................... 56 Results ........................................................................................................................... 59 Descriptive Analyses ................................................................................................. S9 Manipulation Checks ................................................................................................. 64 Preliminary Analyses ................................................................................................ 75 Hypothesis Tests ....................................................................................................... 77 Exploratory Analyses .............................................................................................. 102 Summary ................................................................................................................. 103 Discussion ................................................................................................................... 104 Influences on Intention to Fake ............................................................................... 107 Predictors of Faking Behavior ................................................................................. 110 Effect of Faking on Criterion-Related Validity ....................................................... 116 Implications and Directions for Future Research .................................................... 117 Limitations .............................................................................................................. 121 Conclusion ............................................................................................................... 122 REFERENCES ................................................................................................................ 124 APPENDDC A: Pre-Questionnaire ................................................................................. 133 APPENDIX B: Post-Questionnaire ................................................................................ 136 APPENDD( C: Demographic Questionnaire ................................................................. 138 APPENDDi D: Consent Form ....................................................................................... 139 APPENDD( E: Protocol ................................................................................................. 140 APPENDIX F: Debriefing Form .................................................................................... 146 iv at"; 15*- Us.» In. B\,_ LIST OF TABLES Table 1. Demographic Characteristics of Study Participants ........................................... 47 Table 2. Reliabilities of Personality Scales Across the Honest and Applicant Conditions, and Reliabilities of Difference Scores ................................................................ 51 Table 3. Correlations Among Demographic Variables and Measures. ............................ 60 Table 4. Analysis of Variance Across the Eight Experimental Conditions ..................... 67 Table 5. Means and SDs Across the Eight Between-Subject Conditions ........................ 69 Table 6. Paired Samples t-tests Comparing Means Across the Honest and Applicant Condition ............................................................................................................ 71 Table 7. Regression Analyses Regressing Pre- and Post-Intentions on Attitudes, Subjective Norms, and Perceived Behavioral Control. ...................................... 72 Table 8. Post Intentions Regressed onto Attitude Toward Faking, Subjective Norms, Perceived Behavioral Control, Warning of a lie scale, and Valence. ................ 73 Table 9. Regression Analyses Regressing Post-Intentions on Attitudes, Subjective Norms, Perceived Behavioral Control, and Valence .......................................... 84 Table 10. Regression Analyses with Post Intentions Regressed on Attitudes, Subjective Norms, Perceived Behavioral Control, and Warning of a Lie Scale .................. 85 Table 11. Regression Analyses Regressing Pre-Intentions and Knowledge on Faking Behavior ............................................................................................................. 86 Table 12. Regression Analyses Regressing Post-Intentions and Knowledge on Faking Behavior ............................................................................................................. 88 Table 13. Regression Analyses Regressing Faking Behavior on Pre-Intentions to Fake and the Knowledge Scale ................................................................................... 90 Table 14. Regression Analyses Regressing Faking Behavior on Post-Intentions to Fake and Knowledge ................................................................................................... 92 Table 15. ANOVA Predicting Faking Behavior. ............................................................. 94 LIST OF FIGURES Figure 1. Model of Faking Behavior. ............................................................................... 31 Figure 2. Interaction Between Warning and Valence When Predicting Faking as Measured by the Openness Difference Score ..................................................... 96 Figure 3. Interaction Between Warning and Knowledge on Faking as Operationalized as the Openness Difference Score. ......................................................................... 97 Figure 4. Interaction Between Warning and Knowledge on Faking as Operationalized by the Conscientiousness Difference Score. ........................................................... 98 Figure 5. Supported Relationships. ................................................................................ 105 vi it?» 55.1: me; O . mew Cut? V1521 INTRODUCTION Personality tests are increasingly being used for personnel selection because these tests generally yield moderate validities (Barrick and Mount, 1991; Hough, Eaton, Dunnette, Kamp, & McCloy, 1990; Tett, Jackson, & Rothstein, 1991), yet produce less adverse impact than more traditional selection measures (e. g., cognitive ability tests; Bobko, Roth, & Potosky, 1999). Despite this, there is concern that intentional distortion on the part of applicants may make the results of these tests difficult to interpret because it may result in changes in the test mean, reliability, and validity. However, there is still debate about the effects of faking on the results of personality tests. For example, some researchers have found evidence that faking does not negatively influence the validity of these measures (Abrahams, Neumann, & Githens, 1971; Barrick & Mount, 1996; Cunningham, Wong, Barbee, 1994; Hough et al., 1990; McCrae & Costa, 1983; Ones, Viswesvaran, & Reiss, 1996). However, others disagree with these findings and provide their own evidence that faking does attenuate the validity of non-cognitive measures (Douglas, McDaniel, & Snell, 1996; Dunnette, McCartney, Carlson, & Kirchner, 1962; Kluger, Reilly, & Russell, 1991; Pannone, 1984). There are several reasons that results of research on faking personality and other self-report measures may be inconsistent. An explanation of some of the observed differences between these findings may be that faking is Operationalized in different ways across studies. For example, some researchers have used scores on a lie scale as a measure of faking (e.g., McCrae & Costa, 1983; Hough et al., 1990; Rosse, Stecher, Miller, & Levine, 1998), whereas others have used deviations from the mean (e. g., Dunnette et al., 1962). It is not clear the extent to which each of these measures of faking actually measure intentional distortion (Paulhus, 1986). Second, some of these discrepancies may exist because situational variables are not taken into account. Some situations may inhibit faking on these measures (e. g., knowledge that a lie scale is included on the measure), while other situations may exacerbate the problem (e. g., high competition for a job). Third, a number of the studies that have investigated the relationship between faking and test validity have assumed that there is no variance in faking (e.g., Becker & Colquitt, 1992; Zickar, Rosse, & Levin, 1996). That is, these studies have assumed that everyone who is faking increases their scores the same amount (i.e., by a constant). As Lautenschlager (1994) suggests, if all applicants increase their scores the same extent (i.e., there is no variance in faking), faking will not change rank orders and therefore should have no effect on criterion-related validity. However, if faking does result in changes in the rank orders of applicants (i.e., some applicants increase their scores through faking more than others), then faking has the potential to distort criterion-related validity. Initial research suggests that there is variance in the extent to which individuals fake (Ellingson, Smith, & Sackett, 1999; McDaniel, Douglas, & Snell, 1997; McFarland & Ryan, in press; Rosse et al., 1998). Therefore, the failure to consider this variability in some studies may explain why these studies typically do not find that faking has an effect on validity, while others do. Finally, the discrepancies within the faking literature draw attention to a larger and more fundamental problem. Currently, there is no theory or model of faking behavior. Understanding what contributes to the extent of faking is needed before we can resolve debates about effects due to distortion. Therefore, the present study will use theory to understand an important >4! h". practical problem so that we may better determine when faking may adversely affect personality test results. The present study proposes and tests a model of faking behavior. This model is presented in Figure l. The model incorporates the theory of planned behavior which has been used for several years to predict behaviors as diverse as quitting smoking to stealing (Norman & Conner, 1996; Reinecke, Schmidt, & Ajzen, 1996; Theodorakis, 1992). Specifically, the way in which attitude toward faking, subjective norms about faking, and perceived behavioral control influence the intention to fake will be examined. I will also explore the extent to which the intention to fake predicts faking behavior. Finally, three probable moderators of these relationships will be explored (i.e., valence toward doing well on the test, warning of a lie scale, and knowledge of the constructs assessed by the test). By considering the model of faking presented here, future studies may be able to I consider the relevant issues and influences on faking. Only then can we determine under what circumstances faking does matter. Before describing this model in detail, a definition of faking will first be given, and reasons applicants may fake will be discussed. Second, evidence that applicants can and do fake will be examined. Third, the way in which faking may influence the results of tests will be reviewed. Finally, the model of faking will be described in detail. Definitions and Why People Fake A number of labels have been given to describe the tendency of some individuals to select answers on tests that will result in others viewing them in the most favorable way. This has been referred to as response distortion, social desirability, ,4“ Po; L" r»- “at. 4 Alt. .\u elm. (a [(7‘- fl- \ in faking, impression management, self-enhancement, and intentional distortion (Becker & Colquitt, 1992; Douglas et al., 1996; Mumford & Stokes, 1992, Zerbe & Paulhus, 1987). Paulhus (1984; 1986) suggests that socially desirable responding contains two components: self-deception and impression management. Self-deception refers to the unconscious tendency to see one’s self in a favorable light. Impression management is a conscious attempt to present false information to create a favorable impression on others. Therefore, when applicants taking a selection test consciously do not select the most accurate answer (the one that most closely describes them) but instead select the answer that they believe will make them look most favorable, they are using impression management. In the present study, this type of response distortion, whereby an individual consciously distorts answers in order to be viewed favorably, will be referred to as faking. There are a number of reasons to suspect that people may fake responses on a selection test. Leary and Kowalski (1990) suggest that people are motivated to manage their impressions when the impressions they make are relevant to the fulfillment of their goals and when these goals are greatly valued. Generally, when an applicant takes a selection test, he or she desires to get the job. Individuals may fake in order to obtain a higher score, thus increasing their chances of being hired. Pandey and Rastagi (1979) showed that attempts to manage impressions increased during a job interview when competition for the job became more intense. Therefore, it seems that as the stakes increase, attempts to look the best one possibly can also increase. Other less obvious reasons individuals may fake are suggested by Schlenker and Weigold (1992). They suggest that people regulate the impressions they make to maintain or enhance self-esteem. People generally act to maximize their self-esteem and will influence the content of self-presentation to do so. Answering a test in such a way that others will perceive one positively will help maintain self-esteem. It is clear that there are reasons applicants may attempt to make themselves look favorable. The next questions to be addressed are: Can individuals fake such tests? If they can fake, do they? The following section will review the literature suggesting that applicants can and do fake personality measures. Evidence that Applicants C& and Do Fake For over sixty years research has examined if personality tests can be faked (Bass, 1957; Borislow, 1958; Hunt, 1948; Kelly, Miles, & Terman, 1936; Longstaff, 1948). Findings are very consistent, showing that people can increase scores on such tests when they choose to do so. For example, Dunnette et al. (1962) had 62 salesmen take a forced choice self-description checklist twice in counterbalanced order. In one administration the subjects were asked to answer honestly, whereas in the other one they were told to fake (select the best answers). Subjects in the fake condition had mean scores more than half a standard deviation above those in the honest condition. Since then several more studies have demonstrated similar findings (e.g., Cohen & Lefkowitz, 1974; Hough etal., 1990; Hurtz & Bohon, 1997; Krug, 1978; McFarland & Ryan, in press; Rosse et al., 1998). For example, Douglas et a1. (1996) administered a personality test which contained an agreeableness and a conscientiousness scale to 600 college students. They had a between-subjects design in which half of the subjects were ,4 It! mi 0" l“l -.. fi V <‘ .3 Ltd“ .\l.. l": \a ! randomly placed into the honest condition and the other half were placed into the faking condition (told to make themselves look as good as possible). Results revealed significant group differences of up to .70 standard deviations on the personality scale of agreeableness, and .86 on conscientiousness, with all differences in favor of the fake condition. However, there are some limitations of this research that should be pointed out. Much of the research that has examined the fakability of personality tests has used within subjects designs which require participants to take the test honestly and also under instructions to fake. A problem with such designs is that they generally result in order effects such that individuals who take the test under instructions to respond honestly first increase their scores when asked to fake more than those who did not take the test before being asked to fake (Klein & Owens, 1965). It is not clear in some of these studies if conditions were counterbalanced. Therefore, the results of these studies should be viewed with caution. These studies suggest that responses on personality measures can be faked. When asked to do so, subjects can increase their scores by as much as 1 standard deviation above honest respondents. However, just because applicants can fake personality tests does not mean they do fake. It is important to make the conceptual distinction between fakability and actual faking. If applicants can fake, but do not do so, then responses on a fakable measure may not be biased. Hence, it is crucial to examine the literature for any evidence indicating that applicants actually do fake when responding to personality selection measures. Du conditiom aimed : 3:16. 1 one: 5167771131: the‘l‘fis: Show; Dunnette et al. (1962) compared the scores of those in the fake and honest conditions to the scores of 64 applicants applying for a sales job. The applicants achieved a mean score somewhere between the mean scores of the subjects in the fake and honest conditions, indicating that the applicants faked to some extent. However, to determine whether all or only a few applicants tried to create a good impression on the checklist, a fake key was developed. This was done by weighting those responses showing significant differences between the honest and fake conditions. By examining the fake score distribution it was determined that a cut score of 4 successfully identified 90% of the faked answer sheets at the expense of misclassifying only 9% of the honest ones. Through this investigation it was determined that only a few of the applicants actually faked. Therefore, it was the responses of these few that increased the mean for the applicant group. Pannone (1984) administered a rational biographical questionnaire to over 200 applicants for an electrician position. He included an item which made reference to a piece of electrical equipment that did not exist. Those applicants who said they had used the piece of equipment were considered fakers. It was determined that over one-third of the applicants were fakers and there was a significant mean difference between the scores of fakers and non-fakers such that the mean score of the fakers was over .5 standard deviation units greater than those identified as non-fakers. This study may be criticized in that it is possible that a number of the fakers were not identified. That is, some people who did fake throughout the test did not say that they had operated the fictional piece of equipment. Similarly, those who did not fake throughout the test may have been identified as fakers by their response to the fake item “‘9 .5. @971 Lyn! R.. .- is but were actually just confused by the item or were just careless when responding to the item. Therefore, this one item measure of faking may also be tapping carelessness of responding rather than isolating conscious distortion. Further evidence that applicants may fake selection tests come from Schmit and Ryan (1993). They administered the NEO-FFI, a personality test used to measure the traditional five factor model (extroversion, agreeableness, conscientiousness, emotional stability, and intellectance or openness to experience), to applicants and students. Using confirmatory factor analyses, they found the five factor model fit the data from the student sample but the applicant data contained six factors. The authors suggested this sixth factor (an “ideal-employee factor”) that surfaced in the applicant sample may have resulted from applicant faking. Becker and Colquitt (1992) administered a biodata inventory to 250 real applicants. They also had another group of individuals take the same test under instructions to fake and some under instructs to respond honestly. Results indicate that the scores of the applicants fell between the scores of those in the honest and the fake conditions (they scored lower than the fakers but higher than the honest respondents), indicating that applicants do fake responses. McDaniel et a1. (1997) had 192 subjects complete a conscientiousness and a social influence measure that were mailed to them. Respondents first completed the conscientiousness and social influence measures as they would if they were applying for a job that they really wanted. They were then instructed to respond to these measures a second time honestly. For the conscientiousness measure, 23% of the sample improved their score in the applicant condition and 26% improved their scores in the applicant conditi the san the con their cc Y Scalest “Mining, condition for the social influence measure. However, over 65% of the sample obtained the same score across conditions for the social influence scale as compared to 72% for the conscientiousness measure. When responding as applicants, respondents increased their conscientiousness scale score by .26 standard deviations and their social influence scale score by .16 standard deviations. McDaniel et al.’s (1997) findings may be questioned on the grounds that those who misrepresented themselves may have been less likely to respond to the mail survey for fear that their responses would be traced. However, if this is true then their results probably underestimate the true misrepresentation rate of the sample. This study may also be criticized because the conditions were not balanced; thus, order effects may have been present. Hurtz and Bohon (1997) instructed subjects to answer a personality-based integrity test as they would if they were applicants for a retail sales position. The mean score of those in this applicant condition was over .5 standard deviations higher than those who took the test under instructions to answer honestly. Although there is considerable evidence that applicants do fake responses on personality tests, there are some who do not believe this to be true. Hough et al. (1990) administered a temperament inventory to military incumbents that included a social desirability scale to detect intentional distortion in a favorable direction. All respondents were told to answer the inventory honestly but some were later asked to answer so as to look good (fake condition). The inventory was also administered to 125 individuals going through the Military Entrance Processing Station (MEPS) who were told that the test would be used to make decisions about their careers. Results showed that incum appllC of the fires: v not to.“ been. . incumbents were able to increase scores by faking when told to do so; however, the real applicants did not fake. The social desirability scores of the MEPS were similar to those of the incumbents in the honest condition. One must keep in mind that the MEPS are not traditional applicants. That is, these were individuals who were already enlisted in the military. Therefore, they were not job applicants and may not have been as motivated as job applicants would have been. Additionally, the researchers used scores on a social desirability scale to identify fakers. Research has shown that these lie scales often fail to isolate conscious distortion because they also measure the self—deception component (i.e., the unconscious part) of socially desirable responding (Paulhus, 1986). Not enough information was provided about the social desirability scale used in this study to determine if it was flawed in this manner. Thus, results should be viewed with caution. In summary, most of the evidence suggests that at least some applicants do fake their responses on personality measures. There is also evidence that applicants increase their scores by as much as a half of a standard deviation by doing so (e. g., Dunnette et al., 1962; Stokes, Hogan, & Snell, 1993). Now I will examine why applicant faking may be a concern to those using personality measures for selection purposes. Influence of Faking on the Results of Persorglity Measures There is concern that applicant faking may make it more difficult to interpret personality test results due to its effect on selection decisions, criterion-related validity, and construct validity (i.e., factor structures, scale reliabilities, and correlations with other measures). However, others disagree that faking negatively affects test results. In the following section, the research that has been conducted with respect to the effects of 10 a-- -.‘.U tea: and CU.C the 350.6 the mean “'35 3141: s “dense faking on the results of personality tests will be examined. First I will present the research that has examined how faking may influence selection decisions that are based on personality measures. Second, the possible effects on criterion-related validity will be explored. Finally, the way in which construct validity has been shown to be affected by faking will be examined. Selection Decisions. As discuSsed earlier, research has consistently found that faking increases the mean scores of applicant groups (Becker & Colquitt, 1992; Dunnette et al., 1962; McDaniel et al., 1997; Pannone, 1984). Again, findings indicate that when asked to fake, groups can increase scores by as much as a standard deviation through faking, and applicant groups tend to score a half of a standard deviation greater than groups asked to respond honestly (Hough etal., 1990; Hurtz & Bohon, 1997). Such changes in applicant scores have obvious implications if an organization is using an absolute cut score to select applicants on this type of test. More applicants than would have been selected if everyone responded honestly will be selected if applicants fake. For example, Becker and Colquitt (1992) administered a biodata inventory to over 250 real applicants. To determine the impact of faking on selection decisions, Becker and Colquitt assumed that every applicant distorted his or her score. They then adjusted the score distribution by subtracting three points from everyone’s original score (this was the mean difference between the honest and applicant scores). The hiring decision that was made on the original score (hire or not hired) would stand for 83.1% of the applicants when the new distribution of scores was considered. This study provides evidence that, even if we assume that faking is a constant across individuals (everyone 11 mmgm 11. £1 32:11:. in change sciecwn A commie. the 16PF organizer fakes and increases their scores the same amount through faking), it has the potential to change hiring decisions, such that those who normally would not be hired will be. However, Becker and Colquitt (1992) based this analysis on the assumption that all applicants faked the same amount. Thus, they failed to consider that faking may result in changes in the rank order of applicants. If rank orders change this would likely change selection decisions more drastically. A more thorough analysis of how faking influences selection decisions was conducted by Christiansen, Goffin, Johnston, and Rothstein (1994). They administered the 16PF and a response distortion scale to 495 incumbents in a large factory products organization who were participating in an assessment center. Subjects were told that responses to these measures would be used for future operational purposes (e.g., selection, promotion). They then examined the effect of a social desirability correction on selection decisions for different selection ratios. It was determined that, depending on the selection ratio, correction of scores would have resulted in different hiring decisions than those that would have been made on the basis of uncorrected scores. For example, with a 15% selection ratio 12 candidates who would have been selected on the basis of uncorrected 16PF scores would not have been selected on the basis of faking corrected scores. Results indicated that at lower selection ratios there was a greater percentage of discrepant hires. In fact, they observed a change in rank order for over 85% of the applicants with the application of the social desirability correction to the personality scale SCOTCS . 12 Hough er decrsroris responses was beset increase 1 seiectzon percents; decreased 34? 0." ti. Results u the YEP-.1; c O: 1:6 scale t. 39min} n ”fluke . fake Se} {-1 reg“ DU: mss‘ RI): Ellingson, Sackett, and Hough ( 1999) found similar results. Using the data of Hough et a1. (1990), Ellingson et al. determined the percentage of correct top-down hiring decisions that would be made using the fake responses and then using the corrected responses (i.e., lie scale scores partialled from fake scores). The “true” hiring decision was based on the applicant’s standing in the honest condition. Results indicated that an increase in the selection ratio corresponded to an increase in the proportion of correct selection decisions in the faking condition. Also, holding everything else constant, as the percentage of individuals faking increased, the pr0portion of correct selection decisions decreased. For example, given a 10% selection ratio, if 30% of applicants faked, only 34% of those who would have been hired under the honest condition would be hired. Results were similar for the corrected scores. Therefore, even when faking is corrected, the rank orders are not similar to those in the honest condition. One problem with the majority of these studies is that most have used scores on a lie scale to operationalize intentional distortion. It is not clear how well these scales actually measure faking behavior. However, despite this, there is still considerable evidence to support that there are individual differences in the extent to which individuals fake self-report measures and that this may significantly influence selection decisions (e.g., Dunnette et al., 1962; Hough, 1998; McDaniel et al., 1997; McFarland & Ryan, in press; Rosse et al., 1998). If faking does result in changes in rank orders, it has the potential to alter the criterion-related validity of these measures. Research that has examined how faking influences the criterion-related validity of personality measures will now be reviewed. 13 and ,0 Criterion-Related Vaflm. In their study described earlier, Dunnette et al. (1962) found that faking exerted an effect on the validity of their personality checklist. Managerial ratings were provided for 45 of the salesmen who had answered the checklist under the honest and the fake conditions. Correlations between ratings and checklist scores under the honest condition were relatively large (reaching .38) and significant. However, when the ratings were correlated with scores of those in the fake condition, the correlations decreased and became non-significant. As described earlier, they determined whether all or only a few applicants tried to create a good impression by developing a fake key on the checklist and determined that only a few of the applicants actually faked. Therefore, the authors concluded that even the presence of a few fakers may decrease the validity of such measures. Douglas et al. (1996) obtained job performance ratings from employers of subjects who worked and gave permission to contact their supervisor. Evidence that faking decayed the criterion-related validity of the scales (conscientiousness and agreeableness) was found. The average validity of the scales was .26 for honest subjects and .04 for those who took the tests under instructions to fake. Further analyses revealed results similar to those found by Dunnette et al. (1962). Douglas et al. randomly selected a few of the subjects from the faking condition and added their data to the data of the 97 honest subjects so that the combined sample contained the data of 10 to 25 percent of the faking subjects. Through this procedure they found that as the percent of fakers in the sample increased, the number of fakers scoring within the top ten increased as well. For example, it was found that if 10% of the people faked, five of the top ten subjects were fakers, whereas if 25% of the sample faked, nine of the top ten subjects were fakers. The 14 mean “'1: Moreox'ei validity 0 area. film .\1 their met tor: ‘ient: CKICDI IO ‘ II P.“ his ”013;; rm .. mean validity decreased somewhat as the percentage of fakers in the sample increased. Moreover, it was found that as the number of fakers in the top ten increased, the mean validity of the measures dropped substantially for those who faked. They found that even a few fakers may distort the results of personality measures. McDaniel et al. (1997) had respondents rate how their supervisor would evaluate their overall job performance (on a four point scale) in addition to the measure of conscientiousness, social influence, and misrepresentation. Results indicated that the extent to which respondents misrepresented themselves was negatively correlated with self-reported job performance. They also found that the validities were higher in the honest condition for both the conscientiousness measure (.14 vs. .06) and the social influence measure (.23 vs. .19). A criticism of this study is that the two measures were validated against a self-reported performance measure. However, one would expect that if the self-reported performance measure was not accurate it would be slanted in the positive direction. This would lead to an underestimation of the decline in validity. Therefore, these results may be conservative. Although there is some evidence that faking changes the criterion-related validity of personality measures, there is also evidence to the contrary. Hough et al. (1990) developed a temperament inventory called “Assessment of Background and Life Experiences” or ABLE. This inventory consisted of ten content scales that measured six temperament constructs and a social desirability scale that was included to detect intentional distortion in a favorable direction. Hough et al. administered the ABLE to military incumbents who took the test under both honest and fake conditions. All of these respondents were told to answer questions honestly, but half also took the test under 15 instruct i'aftea' , mentor 1......‘1”- oft DC 93711:» authors I ‘l‘ ¥ . 3M 1~~ . .f-1\ ‘1‘ instructions to fake good and the other half were told to fake bad (possibly so as to not be drafted). The conditions were counterbalanced. Results indicated that the response validity scales did detect the different types of distortion. Moreover, intentional distortion, in a positive direction, did not attenuate the criterion-related validity of the inventory. The method Hough et al. (1990) used to obtain these results is questionable. The data of the incumbents was skewed so that meaningful moderated regressions could not be performed to determine the impact of faking on the inventory. To deal with this the authors took the mean social desirability scale score of those in the fake good experimental condition and used this to divide the concurrent validity sample into two groups. Those scoring at or above the mean of the fake group were labeled “overly desirable,” while those scoring below this mean were labeled “accurate.” They then calculated the criterion-related validities of the ABLE for each of these groups separately. Hough et al. (1990) found that the criterion-related validity did not change substantially for those in the “overly desirable” group; however, the incidence of faking in this group still may not have been very high. As Dunnette et al. (1962) showed, applicants can fake but few actually do. Therefore, it is likely that even those within the top half were not fakers. Instead of splitting the sample in half, Hough et al. should have also examined the top one-third or top ten percent of scorers on the social desirability scale within the incumbent sample. Perhaps if this was done the criterion-related validity would have dropped substantially for the “overly desirable” group. It may also be that the individuals in the “overly desirable” group had true high scores on the measure. 16 dam and con-diner as the pe decision: concigde orders. l1 va‘idzr} o c521} cor Ii : data as H: 31.52:} t he sure if : TS: (”Jimmie fab? Frc P3511101] 0:: Selected 3?! C“wilful-re: Confuse; St ta: {hate u 550185 be,» C As discussed earlier, Ellingson et al. (1999) reanalyzed the Hough et al. (1990) data and found that the corrected mean score was lower than the mean score in the fake condition but greater than the mean score in the honest condition. It was also found that as the percentage of individuals faking increased, the proportion of correct selection decisions decreased. Therefore, using the same data as Hough et al., Ellingson et al. concluded that applicants can fake a personality measure and faking will change rank orders. If rank orders change, faking has the potential to influence the criterion—related validity of these measures. They also demonstrated that the effects of faking are not easily corrected. It should be pointed out that, since the Ellingson et al. (1999) study used the same data as Hough et al. (1990) it is subject to the same criticisms. That is, the Ellingson et al. Study may be flawed because it used a lie scale to assess faking. Therefore, we cannot be sure if the results would hold if an alternative measure of faking was used. The study discussed earlier by Christiansen et a1. (1994) also examined how the criterion-related validity of a personality measure would be affected if corrected for faking. From the 495 participants, 84 were selected for an upper level supervisory position on the basis of the assessment center evaluation. Using the data of the 84 selected applicants for whom criterion data were available, the effect of correction on the criterion-related validities of the traits relevant to the job was examined. Using faking- corrected scores and through using hierarchical regression analyses, it was determined that there was not a significant increase in the criterion variance explained by corrected scores beyond that explained by uncorrected scores. The authors concluded that this 17 demonstr; personalir B. measure 0‘ Responin samples 0' ad self-cl faking or. i \‘oiuntm T: from a m: desrrabm: mltfll 2..“ Tire) eh- d; X, m PEFSOIIE demonstrated that faking does not influence the criterion-related validity of the personality measure. Barrick and Mount ( 1996) administered the Personal Characteristics Inventory (a measure of the FFM of personality) and Paulhus’ Balanced Inventory of Desirable Responding (which measures both self-deception and impression management) to two samples of long-haul semi-truck drivers. They found that both impression management and self-deception related to scores on the personality scales. However, they found that faking on the tests did not significantly reduce their predictivness of supervisor ratings or voluntary turnover. The strongest evidence that faking does not alter criterion-related validity comes from a meta-analysis conducted by Ones et al. (1996). They meta-analyzed the social desirability literature to determine if social desirability operates as a predictor of some criteria, as a mediator, or as a suppressor when personality tests are used as a predictor. They also set out to determine if social desirability is related to real individual differences in personality. They did this by comparing scores on social desirability scales to self and other reports on a personality inventory. The authors found social desirability correlated .37 with emotional stability, .20 with conscientiousness, .14 with agreeableness, and .06 with extroversion. There was no correlation between Openness to experience and social desirability. To determine if these correlations represent the susceptibility of these scales to socially desirable responding or if the correlations represent the relation of social desirability to substantive personality constructs, other people’s (e. g., friends, relatives) ratings of an individual on the personality measures were examined. There were similar correlations between social desirability scale scores and others’ ratings of personality. 18 T335. 11' #350712 pert or: of five tr; \ P‘ P III; Thus, there is evidence that social desirability is related to real individual differences in personality. Ones et al. (1996) also found that social desirability did not predict job performance. They calculated the operational validity of social desirability’s prediction of five frequently used performance criteria (school success, task performance, counterproductive behaviors, training success, and supervisory ratings of job performance). Only training success was predicted by social desirability scale scores. To be a mediator social desirability must correlate with both job performance and personality variables. Because it does not correlate with most performance criteria, social desirability does not act as a mediator. In addition, by partialling social desirability from personality measures, Ones et al. (1996) were able to investigate if such responding impacts the criterion-related validities of the Big Five variables. When this was done, the validities did not change; thus, it was concluded that social desirability did not attenuate the criterion-related validity of the personality measures. It was also determined that social desirability did not act as a suppressor. The authors concluded that the concern with the susceptibility of personality tests to socially desirable responding is needless. Their results indicate that social desirability measures true variance in personality and does not attenuate criterion-related validity. Therefore, controlling for such responding is not necessary. However, Ones et al. (1996) may have overstated their findings because the study contained a major flaw. The scales used by Ones et al. (1996) to measure socially desirable responding not only measure respondents’ conscious attempts to present a positive image (impression management) but also measure the self-deception component of social desirability. For 19 ‘3‘” AI‘LSL '"D" hkua. ‘1 example, it has been shown that the Marlow-Crowne scale and the MMPI K Scale, which were used by Ones et al., load on both the self-deception and impression management factors (Paulhus, 1984; 1986). Therefore, these scales are not only measuring conscious attempts to distort, but also the unconscious component. It is not surprising that measures of self-deception would correlate with personality constructs and not attenuate the criterion-related validity of the personality measures. It has been suggested that if self-deception is controlled for, the usefulness of self-reports may actually decrease because self-deception explains true personality variance (Zerbe & Paulhus, 1987). Therefore, it may very well be that the self-deception component of social desirability was the cause of the correlation between desirability scale scores and personality dimensions. It would be interesting to see if this relationship still existed if self-deception was partialled out of the social desirability scores. Ones et al. (1996) can conclude that the aggregate social desirability scale scores (measuring both self-deception and impression management) used in their analyses correlated with personality dimensions and did not attenuate the validity of personality tests. However, they cannot conclude anything about the relationship between respondents’ conscious attempts to distort their answers to present themselves positively and personality dimensions or validity. Therefore, Ones et al.’s suggestion that faking should be of no concern to those using self-reports of personality is premature. Perhaps self-deception should be of no concern; however, there is still no conclusive evidence which indicates that applicant attempts to consciously distort responses do not bias personality test results. 20 lie or: person; Theon ['81 different. i. .. 4 -, SILDIQCM i shew: 1; Zickar et al. (1996) used a different approach to analyze the effect that faking has on personality measures. They conducted a Monte Carlo simulation using Item Response Theory (IRT) to model faking on a personality test. Results indicated that across different faking magnitudes (variance of faking between respondents) and faking prevalence rates (percentage of fakers in a validation sample), validities did not decrease a significant amount. Also, the mean observed scale difference between validation samples in which fakers were included and samples with no fakers was less than .25 standard deviations for all but the most extreme faking conditions. Thus, faking was not shown to moderate validity. However, their findings indicated that as the percentage of fakers in the sample increased, the percentage of fakers in the top end of the distribution increased as well. Such findings have implications for top-down selection procedures as this indicates that the fakers may be the ones selected. Zickar (1997) sought to replicate the findings of Zickar et al. (1996) but based his analysis on a more complex and realistic model of faking. He examined a computer simulation of faking on the Work Orientation scale from the Army’s Assessment of Background and Life Events (ABLE). In this simulation, Zickar considered three factors not considered by Zickar et al. First, he varied the variance of faking magnitude or the variance between subjects in the extent of faking. Second, the correlation between faking magnitude and the latent trait was systematically manipulated. Finally, the percentage of items deemed fakable was varied. Results indicated that as the variance of faking increased, the validity decreased; however, this correlation was small and not significant. The percentage of items faked had a curvilinear effect on the validity correlation; that is, the most distortion occurred when 75% of the items were fakable and the least amount of 21 distortion occurred when 25% and 100% of the items were fakable. The correlation between the squared percentage of items faked with observed validity was significant (r = -.61), while the correlation between faking magnitude and the latent trait and observed validity was not significant. Overall, validity was influenced the most when the variance in faking was large, there were negative correlations between faking and the latent trait, and 75% of the items were faked. Thus, it seems that in specific situations, faking may decrease the criterion-related validity of non-cognitive measures. Hough ( 1998), in an effort to determine why there are such discrepancies in the literature regarding the relationship between faking and criterion-related validity, sought to determine the cause of these differences. Hough suggested that the different findings were a result of the different samples that were used. By reviewing much of the literature she determined that when incumbents are examined, little faking (as measured by scores on a lie scale) occurs and criterion-related validity is unchanged as a result of this faking. When applicants serve as subjects, more faking (as measured by a lie scale) occurs, therefore the criterion-related validity is slightly compromised, but not substantially. Finally, when research subjects are the focus of the study and are directed to fake (i.e., there is a large amount of faking) criterion-related validity will be substantially reduced. Therefore, Hough concludes that faking in real selection contexts should not be a concern because applicants do not distort responses enough in such situations to reduce the criterion-related validity substantially. There is one major problem with Hough’s (1998) reasoning. She failed to consider that the measure of faking (i.e., the operationalization of faking) may explain these discrepancies between the type of sample used and how faking influences the 22 critenon’ri mcumbent Houever. ' esurratior, nmczec conscious used in sq a ho are 1’! ‘. 13110113 1a criterion-related validity. For example, in all the studies cited by Hough that used incumbent and applicant samples, a lie scale was used as the measure of faking. However, when subjects are directed to fake there is no need to use the lie scale as an estimation of faking because it is already known which group of people faked (i.e., those instructed to do so). Therefore, again, it may be the failure of these lie scales to isolate conscious distortion that accounts for the discrepancies of these results. The lie scales used in studies involving incumbents and applicants may not really be identifying those who are faking. Thus, we cannot know what the true relationship is between faking and criterion-related validity in these samples. Hough (1998) went on to conduct a study that investigated how applicant faking related to criterion-related validity. However, once again, a lie scale was used as the measure of faking. Therefore, it was found that faking did not alter the criterion—related validity of the personality measure in question. Again, we cannot be sure to what extent this scale isolates conscious distortion; therefore, these results are not conclusive. Con_struct Valfity. Studies have also found that faking may influence the construct validity of personality measures through distorting factor structures, the correlations between personalin scales, and the reliability of scales. As discussed earlier, Schmit and Ryan (1993) administered the NEO—FFI to applicants and non-applicants (students). They found that factor intercorrelations were higher for the applicant sample when compared to intercorrelations in the student sample. Moreover, the five factor model fit the data from the student sample, but the applicant data contained six factors. The first factor for the job applicant sample was “a large, work-related personality- characteristic dimension” which they labeled the “ideal-employee factor.” This factor 23 contained most of the items from the Conscientiousness scale, but also included items from the other four scales. One possible explanation for this extra factor is that applicants, unlike students, wish to present themselves as a good employee. It may be that applicant faking created this first factor. A similar study was conducted by Cellar, Miller, Doverspike, and Klawsky (1996). They examined the factor structure of the NEO-PI based on a sample of 423 flight attendant trainees. They also found that the six factor model fit better than the five factor model. However, the sixth factor in their study did not appear to be an ideal employee factor. Instead, this factor appeared to be a method factor. The differences between the findings of Schmit and Ryan (1993) and Cellar et al. may be attributed to the samples used. Schmit and Ryan used actual applicants while Cellar et al.’s sample consisted of trainees who had already passed the selection procedure to become a flight attendant. Therefore, it seems that the sample used by Cellar et al. may have been less motivated to do well or to fake. Cellar et al. also examined applicants for a different type of job. These two influences could have resulted in different factor loadings on the sixth factor. When Ellingson et al. (1999) reanalyzed the Hough et al. (1990) data they also determined the effect of correcting for social desirability on the construct validity of the measure. Ellingson et al. factor analyzed the faked scores, the corrected scores, and the honest scores for the ten personality scales. Results indicated a one factor solution for both the fake and corrected scores, but a two factor (more complex) solution for the honest scores. The authors concluded that faking dissolved the complex 24 multidimensionality of their measure. Thus faking results in scores which no longer reflect personality traits but instead reflect faking behavior. However, Ellingson et al. (1999) Operationalized social desirability by artificially inducing its presence. That is, they specifically asked participants to fake good. Perhaps giving direct instructions to distort responses produced an extremely distorted socially desirable response set consistent with an overpowering social desirability factor. An applied setting may elicit a form of faking that is more refined and may have a different effect (or no effect) on the factor structure of a personality measure. To deal with this limitation of Ellingson et al. (1999), Ellingson, Smith, and Sackett (1999) examined how socially desirable responding influenced the factor structure of personality measures across four different samples. One sample included incumbents, the second examined applicants, the third included applicants and incumbents, and the fourth included applicants, incumbents and students. Additionally, each data set included a different personality measure, and different data collection contexts (i.e., applicants took it within selection context, students within a research context, and incumbents within a developmental context). Within each sample, the fit of the factor structure of the personality test in question was examined separately for those who were labeled “low in social desirability” and those considered “high in social desirability.” Note that individuals were categorized as low and high socially desirable responders according to their scores on a lie scale (those who scored in the top and bottom third of the scale were examined). 25 deception bersonelir a result .3 11'] a deer: Comm;- Results indicated that correlation matrices computed among the personality scales for the high and low social desirability groups were highly similar. Additionally, across all samples the factor structure models demonstrated good fit when applied to the high socially desirable groups. However, socially desirable responding did seem to affect measurement error as error variances were not equal across the two groups (i.e., those high and low in socially desirable responding). However, the authors concluded that faking did not result in significant changes in factor structures. In a study discussed previously, Barrick and Mount (1996) found neither self- deception nor impression management altered the criterion-related validity of a personality test. However, they did find that the correlations observed between the personality scales were higher within their applicant sample as compared to non- applicant samples. They suggest that this increase in the relationship among the scales is a result of applicant faking. Therefore, although they find no support that faking results in a decrease in criterion-related validity, they did provide some evidence that the construct validity of personality measures may be compromised by faking. McCrae and Costa ,( 1983) found evidence that construct validity is not compromised by applicant faking. They administered the NBC inventory and two social desirability scales (the Marlowe-Crowne and the Lie scale from Form A of the Eysenck Personality Inventory) to 215 volunteers. Six months later, the subjects’ spouses were asked to rate their spouse’s personality using the same personality inventory completed by the subject. All correlations between the self and spouse ratings were significant and ranged from .25 to .61. Partialling social desirability out led to significant correlation increases between self and spouse ratings for only two traits (activity and positive 26 emotions . when soci. support t'r. emotions). For most traits, the correlation between the self and spouse ratings dropped when social desirability was partialled out. The authors concluded that these findings support the argument that socially desirable responding does not decrease the validity of personality measures. However, McCrae and Costa only examined socially desirable responding, not intentional faking. The subjects in their study were volunteers so they were most likely not motivated to fake their responses. Also, the two social desirability scales used were not designed to detect respondents’ attempts to consciously distort answers. These measures tap both self-deception and impression management. Therefore, this study may say little about how faking distorts the validity of personality tests. Collins and Glese (1998) found that the Five Factor Model held in an applicant sample. Assuming that the applicants in their sample faked, this is evidence that faking did not influence the construct validity of the personality measure. Ones and Viswesvaran (1998) conducted a meta-analysis to determine the influence of social desirability on the convergent and divergent validity of a personality scale. The number of correlations analyzed exceeded five thousand and the total sample size was over 4 million. Results indicated that both convergent and discriminant validities of the Big Five dimensions remained relatively the same when social desirability was partialled out. The authors conclude that their findings indicate that faking does not adversely influence the construct validity of personality measures. However, again, the meta-analysis included studies that Operationalized faking as scores on a lie scale. Therefore, we cannot be sure if results would be the same had intentional distortion been isolated and partialled out of the correlations. 27 13 personal: faster an expicratt The data and tn 0 respond: shat Sc': 10 distor faking d IOU? Sca xICFaflC Persona Douglas et al. (1996) found faking distorted the construct validity of their personality and biodata scales as measured by multi-trait and multi-method analyses and factor analyses. When the data of those in the honest condition were examined using an exploratory factor analysis, four factors were generated. All four were construct factors. The data from the faking applicants generated four factors but two were method factors and two were construct factors. Thus, a measure based on the responses of honest respondents may have a different structure when respondents fake. This is similar to what Schmit and Ryan (1993) found with the five factor model of personality. Although this study tells us that, when asked to fake, the responses of subjects may distort factor structures, it is not known if the extent to which applicants really fake would be enough to distort the factors in this way. Nonetheless, these results provide evidence that if faking does occur it may change the psychometric properties of the measure. Douglas et al. (1996) also found that intemal consistency reliabilities for their four scales were higher for the faking condition than for the honest condition. McFarland, Ryan, and Ellis (2000) found that when individuals were instructed to fake a personality measure, the reliability of the scales increased. These findings suggest that honest responding may hurt the homogeneity of the items. Subjects who fake make more of an effort to report consistently positive responses. In other words, faking creates an artificial inflation of consistency. Honest applicants report inconsistent behavior because their behavior truly is inconsistent. If it is true, and alphas do change as a result of faking, then these measures are tapping both conscious distortion and the construct which they were designed to measure. This is a violation of the unidimensionality assumption 28 of alpha. This implies that alpha should perhaps not be used as a measure of reliability in settings where some respondents may be faking their answers. Mam There is still debate about the way in which faking influences test decisions. Although there is consistent evidence that applicant faking may alter selection decisions, it is less clear how the criterion-related validity and construct validity of these measures is affected. There are several reasons this area may be plagued by such discrepancies. First, as stated earlier, the operational definition of faking changes across these studies. Some studies use a score on a lie scale as a measure of faking, while others use one item to measure faking, and still others examine increases in scores from an honest condition. Second, it may be that different designs produce different results. Some studies have used within-subject designs, while others have use between subject designs. Third, the samples used may explain these discrepancies. Some studies have used college students and asked them to pretend to be applicants, while others have used applicants or incumbents to determine the effect of faking on validity. Finally, perhaps findings are so discrepant because we still do not understand how various factors may influence the results of each study (e. g., situational cues, knowledge of what the test is measuring). For example, some studies asked participants to fake good but also warned them that a lie scale was included on the measure (Doll, 1971; Hough, 1990), while others have included no such additional instructions (McFarland & Ryan, in press; Pannone, 1984). Some have explained to participants what types of constructs the test in question measures (Cunningham et al., 1994; Dwight & Alli ger, 1997), while many have not done this (Hough, 1998). Therefore, it is not surprising that studies have found such 29 ‘ fi .W W~.I Irilldellgl 011 a per fifiure TClCiIIOn: conflicting results. Before we can really understand how faking may influence test results, we have to understand how these different factors may influence faking behavior. That is, we need a model of faking behavior to understand how and in what way various factors may influence applicant faking. A Model ofjgkfig This study is being conducted to test the model of faking behavior presented in Figure 1. This model is being proposed as a way to predict, explain, and ultimately to understand faking behavior. This model does not address all of the possible explanations for the discrepancies in the faking literature (e.g., study design). However, it does address issues of the operationalization of faking and understanding of the various influences on faking behavior. The current study will examine the usefulness of this model for predicting faking on a personality measure, it may be used to understand faking behavior on any self-report measure (e.g., biodata form, integrity test). Below, each variable in this model and its relationship to faking behavior will be described. 30 £2; coca—32 wEom 833.80 Co omnogocx- 3E 9 3:5... 332.5 mac—am we Eco—2 A Semi 8320m— was; 28m 34 he wEEaB- 3:58:33. L8 coco—w? moose—DEE 3:23.35 33m 3 83:25 wise Beach. 062:2 . L. F 882 0330.33 / 35:00 EoFEem ©0385; 31 of plnne. behavior intenzion ~ Norma : 'he theori Regent's- smoking 10°11. . x] 330' PCIC: intenttor Attitudes, Beliefs, and Perceptions of Control, and Intention to Fake. The theory of planned behavior, originally proposed by Ajzen (1985), will be used to predict faking behavior. This theory has been used extensively within social psychology to predict intentions, and (through intentions) behavior (Ajzen, 1991; Ajzen & Madden, 1986; Norman & Conner, 1996; Reinecke et al., 1996; Schifter & Ajzen, 1985). For example, the theory of planned behavior has been used to predict condom use (Boldero, Moore, & Rosenthal, 1992; Reinecke et al., 1996), recycling (Boldero, 1995), success quitting smoking (Godin, Valois, Lepage, & Desharnais, 1993), and stealing (Beck & Ajzen, 1991). According to this theory, one’s attitude toward the behavior, subjective norms, and perceived behavioral control predict the intention to perform the behavior. In turn, intentions predict behavior. Attitude toward the behavior refers to the degree to which a person has a favorable or unfavorable evaluation of the behavior in question. More specifically, it is an evaluative appraisal of the consequences of acting or not acting (Ajzen, 1991). This appraisal may be either favorable or unfavorable, and is generally expressed with semantic differential items (e.g., good-bad, pleasant-unpleasant). Subjective norm is a social factor predicting intentions. This refers to the perceived social pressure to perform or to not perform the behavior. For example, items used to measure subjective norms assess the likelihood that important individuals (e. g., parents, friends) would disapprove or approve of the behavior. Therefore, an individual who thinks friends and family members who are important to them would disapprove of faking on a personality test will have a negative subjective norm for faking behavior. 32 51' C) n (D ‘.’ '7 I". Perceived behavioral control is the third factor proposed to predict intentions. This refers to the individual’s belief regarding the ease or difficulty with which the behavior can be performed. Individuals who believe they can perform the behavior are more likely to intend to behave in that manner. The theory of planned behavior is an extension of the theory of reasoned action (Fishbein & Ajzen, 1975). Unlike the theory of reasoned action, the theory of planned behavior does not assume that the behavior in question is completely under one’s volitional control. Therefore, the theory of planned behavior goes a step further than the theory of reasoned action by considering an individual’s perception of control over the behavior in question. It is believed that the inclusion of this additional variable is why the theory of planned behavior has been shown to predict behavior better than the theory of reasoned action (Ajzen, 1991). The theory of planned behavior will be used to predict the intention to fake in the following way. Specifically, it is hypothesized that attitudes toward faking personality tests in selection contexts, subjective norms of faking (i.e., perceptions of whether or not important others would approve or disapprove of faking behavior), and perceptions of whether or not one could fake the test, will have a direct effect on the intention to fake. It has been suggested that attitudes regarding dishonesty (whether it be theft, cheating, lying, etc.) are related to intentions to behave in a dishonest manner. For example, it has consistently been found that positive attitudes toward a behavior lead to greater intentions to perform that behavior (e.g., Ajzen, 1991; Boldero, 1995; Boldero et al., 1992). Beck and Ajzen (1991) found that attitudes toward dishonest actions predicted intentions to cheat on a test, shoplift, and lie. Since positive attitudes toward a behavior 33, have con: following Hmothe fake. or acre; instance that mo. based 0 other Ct society 0"] A; have consistently been linked to stronger intentions to perform that behavior, the following prediction will be made: Hypothesis 1: Attitude toward faking will be positively correlated with the intention to fake. It has been suggested that the belief that certain dishonest behaviors are common or acceptable is a determinant of many forms of dishonesty (Murphy, 1993). For instance, it has been found that an individual is more likely to steal if he or she believes that most people steal (O’Bannon, Goldinger, & Appleby, 1989). Most honesty tests are based on evidence that suggests that individuals who are more likely to commit thefts and other counterproductive acts are more likely to believe that such behavior is accepted by society (Hollinger, 1989). Research on the theory of planned behavior has consistently found that when subjective norms toward a behavior are positive, the intention to perform that behavior will be greater (Ajzen, 1991; Beale & Manstead, 1991; Schifter & Ajzen, 1985). This means that when individuals perceive that important others would approve or encourage the behavior in question, they are more likely to intend to engage in that behavior. For example, Beck and Ajzen (1991) found that favorable subjective norms of cheating, shoplifting, and lying were positively related to intentions to behave in these ways. That is, individuals who thought their friends and family would approve of such behavior (or not disapprove) would be more likely to intend to engage in these behaviors. It is expected that these results will generalize to faking on a selection test. Hypothesis 2: Subjective norms toward faking will be positively related to the intention to fake. 34 behavlt‘r behavior suggest: Hos ex e predict 1' Se l {-6 iii Perceived behavioral control is the third variable in the theory of planned behavior that is hypothesized to influence intentions. Ajzen (1991) suggests perceived behavioral control refers to an individual’s perception of the ease or difficulty of performing the behavior of interest. As Ajzen points out, this definition is very similar to Bandura’s perceived self-efficacy (Bandura, 1986; 1991). He suggests that the theory of planned behavior places this construct (self-efficacy) within a more general framework of the relations among beliefs, attitudes, intentions, and behavior. Therefore, Ajzen suggests self-efficacy is very similar to, if not the same as, perceived behavioral control. However, since the model in this paper is integrating the theory of planned behavior to predict faking behavior, the term perceived behavioral control will be used instead of self-efficacy because this is the term used within the theory’s framework. Research has consistently found that perceived behavioral control is positively related to intentions such that when individuals perceive that they have control over the behavior in question, the intention to perform the behavior is greater (Ajzen, 1991; Beale & Manstead, 1991; Schifter & Ajzen, 1985). For example, Beck and Ajzen (1991) found that perceptions of behavioral control were highly and positively correlated with intentions such that those who believed they had control over cheating, lying, and shoplifting were more likely to intend to behave in these ways. Similar results are expected in this study, such that individuals who perceive control over faking behavior (i.e., those who think they could fake a personality selection test) will have greater intentions to fake. Hypothesis 3: Perceived behavioral control of faking will be positively correlated with the intention to fake. 35 intent-0 is. the g the bet. Mansre. Ajzen. i from 10. Thereft: tntentror 31036 “j Intention to Fake and Faking Behavior. It has consistently been found that the intention to engage in a behavior is related to performing the behavior in question. That is, the greater the intention to perform a certain behavior, the greater the likelihood that the behavior will be performed (Ajzen, 1991; Ajzen & Fishbein, 1980; Beale & Manstead, 1991; Beck & Ajzen, 1991; Boldero, 1995; Boldero etal., 1992; Fishbein & Ajzen, 1975; Schifter & Ajzen, 1985). Intentions have been shown to predict everything from losing weight (Schifter & Ajzen, 1985) to stealing (Beck & Ajzen, 1991). Therefore, in the present study it is expected that a similar relationship between the intention to fake and faking behavior will be found. Hypothesis 4: The intention to fake will be positively related to faking behavior such that those with a greater intention to fake will fake to a greater extent. Although the relationship between intentions and behavior is strong, it is far from perfect. This indicates that there are other variables at work that are influencing this relationship. The same is true of the relationship between intention to fake and faking behavior. The faking literature suggests some moderators of these relationships. Therefore, the model of faking presented in this paper integrates the theory of planned behavior with some variables that have empirically been shown to affect faking behavior. Specifically, it is expected that situational factors will moderate the relationships between attitudes toward faking, subjective norms toward faking, and perceived behavioral control and the intention to fake (see Figure 1). Additionally, ability to fake (i.e., knowledge of the constructs being measured) will moderate the relationship between the intention to fake and faking behavior. A detailed discussion of these moderators will now be presented. 36 orrna)' Moderators. Although the theory of planned behavior has been successful at predicting intentions and behavior, some have suggested that the theory would explain more variance if moderators were considered in the framework (Bagozzi, 1992). Therefore, other predictors of the intention to fake and faking behavior will be examined within the context of the theory of planned behavior. Research in the faking literature has found that there are multiple situational factors that may influence faking behavior in applied settings. These situational factors may be situations the applicant is placed in with respect to his or her life circumstances, or may be more directly related to the selection context (e.g., valence for a job, concern with getting caught faking). For example, it has been suggested that when valued outcomes are at stake, applicants will fake more (Mabe & West, 1982). Therefore, an individual in a financial crisis may really need the job in question such that she would fake to ensure she gets the job. It has also been shown that the more desirable a job, the greater the likelihood that the individual will fake responses on a selection test (Fletcher, 1990). Therefore, there is evidence that valence for a job may predict faking behavior. In general, it has been suggested that the more desirable the outcome is (doing well on a test), the more likely an individual is to distort his or her responses (Schlenker & Weigold, 1992). In the present study, although valence for a specific job will not be measured (as this is a lab study), the valence one has for the outcome of doing well on the test will be examined. Specifically, a monetary incentive will be offered to some participants, such that those who perform well on the test will receive money (this will be explained in more detail below). 37 annude ofnotd 0065 DC thecon- Skater Sublet: have a H\ As Figure 1 shows, situational factors will moderate the relationship between attitudes toward faking, subjective norms toward faking, and perceptions of behavioral control and the intention to fake. It is expected that even if an individual has a negative attitude toward faking (or negative subjective norm toward faking), the individual may still intend to fake if the outcomes of doing so are very desirable, or if the consequences of not doing so are negative (e.g., not obtaining a job). Likewise, even if an individual does not think he or she can fake a personality test, he or she may still intend to fake if the consequences of doing so are very positive. In the present study it is predicted that those who are offered an incentive will have more positive valence for doing well on the test. Based on the above discussion, the following predictions will be made: Hypothesis 5a: Valence for doing well on the test will moderate the relationship between attitudes toward faking and intention to fake such that those with high valence will have a greater intention to fake, regardless of attitudes toward faking. Hypothesis 5b: Valence for doing well on the test will moderate the relationship between subjective norms for faking and intention to fake such that those with high valence will have a greater intention to fake, regardless of their subjective norm of faking. Hyp_othesis 5c: Valence for doing well on the test will moderate the relationship between perceptions of behavioral control toward faking and the intention to fake such that those who have high valence will have a greater intention to fake, regardless of their perceptions of behavioral control. Characteristics of the specific selection situation may make applicants more or less likely to fake. For instance, it has been suggested that concern with getting caught faking may decrease faking behavior on selection tests. Research has consistently found 38 vent-la? that Mi: , . apph.ar that in self-report measures, verifiable and objective items are faked less than items that are unverifiable and subjective (Becker & Colquitt, 1992; Shaffer, Saunders, & Owens, 1986; Stokes et al., 1993). For example, Stokes et al. found that applicants faked verifiable items less than items measuring personal preferences. Research has also found that when applicants are told their test answers will be verified, they fake less than applicants who are not given such information (Hough et al., 1990; Mael, 1991). It has been suggested that this decrease in applicant faking (as measured by both lie scale scores and mean differences from honest groups on the test) on verifiable and objective items (and when told their responses will be verified) is because applicants fear that their attempts to fake will be detected. If applicants are identified as fakers, they fear this would result in decreasing their chances for the job (Lautenschlager, 1994). As faking is generally done to increase one’s chances of getting a job, it seems that concern with being caught, and thus being eliminated from the selection process, would decrease faking behavior. When using personality tests, it is less feasible to verify answers or to create objective items (as compared to many biodata forms or integrity tests) because whether or not an individual has desirable personality characteristics is being measured. A measure of this type is inherently less verifiable and more subjective (Wheeler, Hamill, & Tippins, 1996). Therefore, warning applicants that their responses may be verified (in order to reduce faking) may not work as it will be obvious to the applicants that such verification would be difficult, if not impossible, to obtain. Therefore, other means have been used to deter faking on personality tests and other less verifiable self-report measures. 39 chore. lllfflai. 1., Altlflt‘x', 1111“,}: junior hill“ 01“ l mini: “fife “ Research has shown that telling individuals that a lie scale is included on the test that will identify those who are faking responses may decrease faking (Doll, 1971; Kluger & Colella, 1993; N ias, 1972; Schrader & Osbum, 1977; Wheeler et al., 1996). To determine if warning that faking attempts may be detected decreased faking on a test of mental incompetence, Hiscock, Layman, and Hiscock ( 1994) administered two forced choice tests — a knowledge test and moral reasoning test - to 105 inmates. A third of the inmates were instructed to take the tests honestly, a second third were told to fake bad, and the last third were told to fake bad but also that there were ways that the inmates may be identified as fakers (e.g., by appearing excessively impaired or answering difficult items correctly and easier ones incorrectly). Results indicated that those asked to fake without being warned scored significantly lower on the knowledge test than those who were told to fake but given a warning. Both groups scored lower than those responding honestly. For the moral reasoning test, although not statistically significant, those in the unwarned fake condition scored lower than those in the warned fake condition. Similarly, Nias (1972) had 262 twelve and thirteen year old children take the junior Eysenck Personality Inventory. All children were told to take the test honestly, but half of them were also warned that a lie scale was included in the measure to detect individuals who may not be responding honestly. Results indicated that the children who were warned of the presence of a lie scale scored significantly lower on the lie scale and had higher mean scores on psychoticism and neuroticism. Nias also examined if the warning reduced faking more than if one just corrected scores for such responding. To test this, those who had high lie scale scores within the group that was not warned of a lie scale were taken out of the sample and the means for the various scales were recalculated 4O with those peoPle excluded. Results indicated that mean scores on psychoticism and neuroticism were still much lower compared to the means for the group that was warned a lie scale was included on the measure. Therefore, it seems that warning may be more effective at dealing with faking than correcting scores for faking. Doll (1971) had 300 college graduates receiving armed service training take a biodata instrument under instructions to respond honestly. Some of the subjects were then told to fake to look good but to be prepared to defend their answers, another group was told to fake to look good but to be aware that a lie scale may be included to detect faking, and the last group was told to fake to look as good as possible. Results indicated that subjects could increase their scores by faking and that the largest score increase was in the “fake to look as good as possible” condition while the least amount of increase was seen by those in the condition which instructed subjects to be aware of a lie scale. Schrader and Osbum (1977) examined how warning that a lie scale was included on a biodata form influenced faking. They asked participants to fake a biodata form. However, before taking the biodata form, half of the participants were told a lie scale was included in the measure to identify those who were giving false responses. It was found that the group warned that a lie scale may have been included in the measure scored a half of a standard deviation lower than those who were not warned. Recently, Wheeler et al. (1996) replicated these results, finding applicants who were warned a lie scale may be included on a personality measure scored lower on the lie scale and had lower mean scores then applicants not given the warning. It is believed that this occurs because applicants do not want to be identified as liars and thus decrease their chances of being selected for the job (Kluger & Colella, 1993). 41 Therefore, it seems that warning that a lie scale is included in the measure may decrease the intention to fake, even if subjective norms, attitude toward faking, and perceptions of behavioral control are high. Therefore, the following hypotheses will be made: Hypothesis 6a: Warning that a lie scale is included in the measure will moderate the relationship between attitude toward faking and the intention to fake such that those warned that a lie scale is included in the measure will intend to fake less, regardless of their attitude toward faking. Hypothesis 6b: Warning that a lie scale is included in the measure will moderate the relationship between subjective norm toward faking and the intention to fake such that those warned that a lie scale is included in the measure will intend to fake less, regardless of their subjective norm toward faking. Hypothesis 6c: Warning that a lie scale is included in the measure will moderate the relationship between perceptions of behavioral control over faking and the intention to fake such that those warned that a lie scale is included in the measure will intend to fake less, regardless of their perceived behavioral control. Another moderator considered in the model of faking behavior is the ability to fake. The ability to fake refers to how successful an individual is at increasing his or her score through faking, when he or she chooses to fake. An individual may intend to distort responses and attempt to do so, but not have the resources available to successfully increase his score through distortion. That is, he may not have the ability to distort. There are several factors that may increase the ability of an individual to fake self-report measures. One that has been well researched, and will be examined in this 42 study, is knowledge of the construct being measured (Cunningham et al., 1994; Dwight & Alliger, 1997; Holden & Jackson, 1981; Napier, 1979). Goldman and Olczak ( 1976) examined the effect of knowledge about self- actualization on faking the Personal Orientation Inventory (a measure of self- actualization). All participants first took the inventory under instructions to respond honestly. Then, half of the participants were given a class that described and discussed self-actualization. After this, one-third of the participants were told to take the inventory again but were told to fake bad, one third were told to fake good, and the last third were told to take the test honestly again (in each condition half of the participants had been given information about self-actualization while the other half had not). Results indicated that both groups (those not given information about self-actualization and those given the information) asked to fake good were able to significantly increase their scores and both groups were also able to decrease their scores by a significant amount when asked to fake bad. However, those given knowledge of self-actualization were able to increase their scores more than those not given this information when asked to fake good and decreased their scores significantly more than those not given this information when asked to fake bad. Thus, when provided with information about the construct the test was measuring, participants were able to fake to a greater extent. Cunningham et al. (1994) explored the effects on test scores of providing specific content information about the major constructs involved in the Reid Report, a popular overt integrity test. The authors instructed all participants to respond to the measure as they would if they were taking the test as a job applicant. However, some participants were given information about the punitive scale (what it was trying to measure and how it 43 din: r. -~ receix : ' desist. on the measured it), some about the projective scale, and some about both dimensions of the test, while another group was given no information about any test dimensions. In an attempt to motivate subjects, they were told that those scoring within the top five percent would receive ten dollars. Results indicated that those who scored lowest on the test were subjects who received no information about what the test measured. Those given information on a dimension of the test had significantly higher mean scores on that dimension as compared to those who did not receive this information. The groups receiving information on both dimensions had the highest mean total scores (.5 standard deviations greater than the control group). However, there were some generalized effects on the non—targeted dimensions such that information on one dimension increased the mean score on the other dimension slightly. These results suggest that applicants can improve scores through faking if they are provided with information about the test. These findings may be exaggerated given that students served as subjects. Real applicants may give more thought (due to greater motivation) to what the test is trying to measure and not need to get information on the nature of the test because they would be motivated to figure it out on their own. However, such information may have helped students who have less experience with selection tests and who are less motivated to increase scores through faking. Nonetheless, this study indicates that the ability to figure out what the test is measuring may be a good predictor of faking. Dwight and Alliger (1997) replicated the findings of Cunningham et al. (1994). They had 120 undergraduates take the Employee Integrity Index (EH) which is an overt integrity measure developed by Ryan and Sackett (1987). Subjects were placed in one of three conditions: fake good, honest, or a coaching condition. Those in the coaching condr respe.’ of the than :1 the is. those donor 111C763 condition were told what integrity tests are designed to measure and taught how to respond to such items so as to obtain a high score. Results indicated that the mean score of the coaching condition was significantly higher (by almost .5 standard deviation units) than the mean score of those in the fake good and honest conditions. The mean score of the fake good group was also significantly greater (by over one standard deviation) than those in the honest condition. Cunningham et al. (1994) and Dwight and Alliger (1997) demonstrated that coaching individuals about the nature of the construct being measured increased individuals’ ability to increase their scores. It should be noted that no research could be found that has examined knowledge of constructs and faking on personality measures. There is no reason why the results pertaining to integrity tests should not also apply to personality testing. Both are self- report measures. Also, although not empirically tested, it has been suggested that applicants will be more successful at increasing their scores on personality measures through faking if they can tell what the test is measuring (Anastasi & Urbina, 1997). Therefore, it is predicted that knowledge of the constructs that the test measures will moderate the relationship between the intention to fake and faking behavior. Again, an individual may intend to distort responses, but not have the necessary ability to do so. Hymthesis 7: Knowledge of constructs measured will moderate the relationship between the intention to fake and faking behavior such that those who are given information about the constructs measured on the personality test and have a high intention to fake the test will increase scores more than those with similar intentions but not given this information. 45 COUt ling. It should be noted that the present study will operationalize faking in two ways: scores on a lie scale and difference scores (i.e., change in scores from the honest to applicant condition). As discussed previously, one of the reasons findings may be so discrepant in this literature is a failure to have one operationalization of faking across studies. Some studies use a difference score to measure faking, while others use lie scale scores. To determine if the results are consistent across different measures of faking, the present study will explore the effects on faking and the effects of faking, using both a lie scale and difference scores. Faking and Validity. As discussed, one of the main reasons faking is even a concern to those using selection tests is the potential impact faking may have on the validity of personality measures. Therefore, in addition to testing the hypotheses above, the extent to which faking influences the validity of the personality measure will also be examined using GPA as the criterion measure. Research has demonstrated that conscientiousness is a valid predictor of job performance. Additionally, conscientiousness has been shown to predict educational criteria such as GPA. Therefore, to determine if the validity of the test changes as a result of faking, GPA will be correlated with conscientiousness test scores from both the honest and the applicant conditions. If the correlations change substantially across conditions, we can be fairly certain that faking is changing the validity. However, given that the empirical results are mixed with regards to how faking may influence the results of such tests, no hypotheses will be made, but the issue will be explored. 46 Method m Participants were 547 undergraduates from a large university in the Midwest. This sample size results in considerable power to detect even small effects (Cohen, 1988; 1992). Table 1 contains specific information about the demographic makeup of the participants. As the table shows, the majority of the participants were White females under the age of 20. Table 1. Demographic Characteristics of Study Participants Total Sample Size 570 Sex Male 171 (30%) Female 376 (66%) Race (’Whitc 426 (74.7%) 1 African American 45 (7.9%) Hispanic 11 (1.9%) Asian 40 (7.0%) Other 21 (3.7%) Age 19.80 (2.96) Past Experience With Personality Tests Yes 405 (71.2%) No 43 (7.5%) Design A 2 (incentive vs. no incentive) x 2 (warning of lie scale vs. no warning) x 2 (knowledge of constructs vs. no knowledge) between-subjects design was used. For the first factor, participants were asked to respond to a personality measure as if they were an applicant for a job they would like to have. However, approximately half of the 47 participants (11 = 253) were also given an incentive such that they were told that those who scored in the top 15% would receive $20 while the other participants (11 = 294) were not offered this incentive. Other studies have shown that $20 is seen as very desirable to undergraduates (McFarland & Ryan, in press; McFarland et al., 2000). Therefore, it is reasonable to expect that those offered this incentive would have significantly higher valence toward doing well on the test than those not offered the incentive. The second factor was warning of a lie scale. Approximately half of the participants (p = 285) were told that a lie scale was included in the measure so that fakers could be identified, while the other half (p = 262) were not told a lie scale was included in the measure. The third factor was knowledge of constructs being measured. Half of the participants (n = 276) were told what constructs the personality test measured and these constructs were defined for them prior to taking the personality test. The other half of participants was not given this information (13 = 271). It should be pointed out that in addition to taking the personality test under instructions to respond like an applicant, participants also took the personality test under instructions to respond honestly. These honest responses were only collected so that difference scores could be created (these difference scores serve as one measure of faking behavior). However, the honest and applicant conditions were counter-balanced to determine if the order of the conditions effected results. Pre-guestionnaire. Approximately one week before attending the experiment, participants took the pre-experiment questionnaire (as a pre-questionnaire on the Internet). This measure included items that measured participants’ attitudes toward faking on selection tests, subjective norms for faking, perceptions of behavioral control 48 over faking, and intentions to fake in the future on selection tests (pro-intentions to fake). In the study conducted by Beck and Ajzen ( 1991), participants were asked to respond to similar items (e.g., items that asked about attitudes, subjective norms, and perceived behavioral control toward stealing and cheating). The results indicated that individuals were very willing to indicate what would be perceived as undesirable responses (e. g., positive attitudes toward theft). However, to increase the likelihood of honest responses to this measure, participants were told that their responses would be kept completely anonymous and that the answers they provided would not affect them in any way. Additionally, they were encouraged to respond honestly. All items were adapted from Beck and Ajzen who found adequate reliabilities for the above measures (all greater than .65). See Appendix A for a reproduction of these measures. Attitude Toward Faking. Participants indicated their attitudes toward faking a selection test on five, 5-point semantic differential-type response scales: good-bad, pleasant-unpleasant, foolish-wise, useful-useless, unattractive-attractive. The internal consistency of this scale was .86. Subjective Norms. Four items were used to measure subjective norms toward faking. These items were answered on a 5-point scale from strongly disagree to strongly agree. An example of an item on this scale is “Most people who are important to me would look down on me if I lied on a selection test.” The internal consistency of this scale was .73. 49 Perceived Behavioral Control. Three items were used to measure perceived behavioral control over faking. These items were on a 5-point scale from strongly disagree to strongly agree. An example of an item on this scale is “It would be easy for me to lie on a selection test.” The internal consistency of this scale was .79. Pre-Intentions to Fake. Generally, intention measures are given prior to the performance of the behavior that is to be examined (Ajzen, 1991). Therefore, ten items were used to measure intentions to fake in the future on selection tests. Items were on a 5-point scale from strongly disagree to strongly agree. An example of an item on this scale is “I would NEVER lie on a selection test.” The internal consistency of this scale was .89. Personality Test. The NEO-FFI (Costa & McCrae, 1989), a measure of the five factor model of personality, was used to assess personality. This measure is a short version of Costa and McCrae’s (1985) NEO-PI. The five factors measured are neuroticism, extroversion, openness to experience, agreeableness, and conscientiousness. Each of the five factors are measured by 12 items with a Likert response format ranging from 1 (strongly disagree) to 5 (strongly agree). The internal consistencies for each of the scales across both the honest and applicant conditions are reported in Table 2. 50 Table 2. Reliabilities of Personality Scales Across the Honest and Applicant Conditions. and Reliabilities of Difference Scores Scale Honest Applicant Difference Score Neuroticism .88 .88 .61 Extroversion .84 .84 .42 Openness .76 .72 .21 Agreeableness .77 .78 .24 Conscientiousness .86 .88 .75 51 Post-Questionnaire A post questionnaire was handed out after the participants took the personality test under the applicant condition. Most of the items were used as manipulation checks (i.e., to assess if the manipulations were salient to participants). Intentions to fake the personality measure were also assessed. See Appendix B for a reproduction of this questionnaire. Valence Toward Doing Well on the Test. This measure assessed how much the participants valued the outcome of doing well on the test. It was anticipated that those who were offered the incentive for doing well on the test would have higher valence toward doing well on the test than those not offered the incentive. Thus, this measure served as a manipulation check to determine if those who were offered the incentive felt more positively toward doing well. Five items were included on this measure. Responses to all items were made on a Likert type response scale ranging from 1 (strongly disagree) to 5 (strongly agree). An example of an item on this measure is “doing well on the test is very desirable to me.” The internal consistency of this scale was .90. Concern With Being Caught Faking. This measure served as a manipulation check. It was used to estimate how concerned individuals were that they would be identified as a faker and whether or not they believed such faking attempts would be detected by the scale. It was expected that those who were warned about the lie scale being included on the test would have greater concern that their false responses would be detected by the lie scale. This measure contained nine items. These items had a Likert type response format ranging from 1 (strongly disagree) to 5 (strongly agree). An 52 example of an item included on this scale is, “I was concerned that the lie scale would identify me as a liar.” The internal consistency of this scale was .77. Knowledge of Constructs Being Measured. Eight items were used to measure knowledge of constructs being measured. This measure served as a manipulation check to determine if those who were given information about the constructs the personality test was measuring were better able to determine which items related to the constructs. This measure contained items that asked participants to indicate if they could successfully identify the constructs that were being assessed on the test. Items were on a 5-point scale from strongly disagree to strongly agree. An example of an item on this scale is, “Some items measured how reliable and responsible a person is.” The internal consistency of this scale was .80. Post-Intention To Fake. A questionnaire assessing the extent to which participants intended to fake the personality test under the applicant condition was administered. As stated earlier, intention measures are generally given prior to the performance of the behavior that is to be examined (Ajzen, 1991). This is why the present study administered an “intentions to fake in the future” measure on the pre- questionnaire. However, this pre-intention to fake measure was administered about a week before the participants actually took part in the experiment. It has been suggested that, for it to predict behavior, intentions should be measured as close to the time at which the individual has the opportunity to perform the behavior in question (i.e., faking; Ajzen, 1991). In this study, if participants were asked their intentions to fake immediately prior to the administration of the test, it might have influenced their responses on the test (i.e., some people may have never even thought about faking, but since the experimenter asked 53 about faking right before the test, some participants may have been likely to fake; Feldman & Lynch, 1988). Also, the design is such that post-intentions must be measured to test two of the moderation effects (valence and lie scale). That is, an intention to fake measure must be administered after participants are offered the incentive and after they have been told that a lie scale is included on the test. Otherwise, it cannot be determined if these manipulations influenced intentions to fake. For these reasons, intention to fake was also measured immediately following the test administration. However, just because the post-intention measure was given after the test does not mean that this is simply a measure of faking behavior on the test. Post-intention items asked if individuals intended to fake the test. Some individuals may have intended to do so, but may not have been successful. Therefore, other measures of faking behavior (i.e., difference scores and impression management scale scores) serve as the actual measures of the extent of faking. This post-intention measure included seven items that used a Likert type response format ranging from 1 (strongly disagree) to 5 (strongly agree). An example of an item on this scale is, “I intended to lie on this test.” The internal consistency of this scale was .91. Demographic Questionnaire. A questionnaire assessing various participant demographics was also included and assessed factors such as race, sex, years in college, and GPA. This measure also had participants indicate the amount of experience they have had with personality tests. This measure also asked participants whether or not they were thinking of any specific job for which they were applying when they took the personality test under instructions to behave like an applicant. If so, they were asked to 54 indicate what kind of job they imagined they were applying for. See Appendix C for a reproduction of this questionnaire. Faking Behavior. Across the faking literature, two operationalizations of faking are most often used: lie scale scores and difference scores. Therefore, in this study, faking behavior was measured in these two ways. First, within the NEO-FFI, items from Paulhus’ (1984; 1991) BIDR scale were included. This is a lie scale that is used to measure both self-deception (unconscious distortion) and impression management (conscious distortion). The self-deception scale contained 20 items. An example of an item on this scale is “I have not always been honest with myself.” The internal consistency of the self-deception scale was .74 in the honest condition and .86 in the applicant condition. The impression management scale also contained 20 items. Scores on the Impression Management scale were used to measure intentional distortion (i.e., faking; Paulhus, 1986; 1991). An example of an item on this scale is “When I hear people talking privately, I avoid listening.” The internal consistency of the impression management scale was .76 in the honest condition and .90 in the applicant condition. Although both self-deception and impression management are measured, we are most concerned with impression management scores because this is the measure used to assess intentional distortion. Faking behavior was also assessed by examining differences in scale scores across the honest and applicant conditions. Several studies have used changes in test scores from honest to applicant or fake conditions to operationalize faking (i.e., the difference between the two administrations is considered to be due to faking; e. g., Doll, 55 1971; Ellingson et al., 1999; McDaniel et al., 1997; McFarland & Ryan, in press). Therefore, a measure of faking behavior (in addition to the impression management scores) was derived for each of the personality scales by subtracting the score received in the honest condition from the score received in the applicant condition. This resulted in 5 difference scores for each individual. These difference scores were used as one measure of faking behavior. By using the difference score we can be more confident that intentional distortion (i.e., faking) is being isolated. Self-deception may occur in both the honest and applicant conditions, but as described earlier, self-deception remains constant across situations (Paulhus, 1986) and so should not influence the difference score. Table 2 contains the reliability of the difference scores. The following formula was used to calculate the reliability of the difference scores: rdd = (O'd2 - Ocdz) \ 0,12; where God: = 01.2 (l - rm.) + of (1 - ram); with h representing the measure in the honest condition, a for the applicant condition, and of representing the variance of the difference score (Rogosa, Brandt, & Zimowski, 1982; Tisak & Smith, 1994). Only two of the difference scores achieved adequate reliabilities (neuroticism and conscientiousness). Thus, further analyses that include the other three difference scores (agreeableness, openness, and extroversion), should be viewed with caution. Procedure Participants signed-up on-line via the Internet to participate in the experiment. They were directed to an Internet site and could select to participate in a number of experiments for credit. Before participants could sign-up to participate in the present study, they were required to fill out a 22 item questionnaire. This was the pre- questionnaire containing the attitude toward faking, subjective norms toward faking, 56 perceptions of behavioral control, and pre-intention to fake measures. Therefore, on average, this survey was taken one week before the participants came to the experiment. The experiment session began by having participants read and sign an informed consent form (see Appendix D for a reproduction of the consent form). Half of the participants were randomly assigned to take these tests under the honest condition first and the other half were asked to complete the tests under the applicant condition first. In the honest condition participants were given the following instructions: Please answer the following questions on the personality test as honestly as possible. Your answers will remain completely confidential. Your answers will be used for research purposes only, and will not be used to evaluate you in any way. The nature of the project requires that you answer the following questions as honestly as possible so please provide as accurate answers as you can. Half of the participants were also told the five constructs the personality test was created to measure. A definition of each construct was also provided (see the protocol in Appendix E for specific instructions). 1 Before taking the personality test within the applicant condition, all participants were given the following instructions: When answering the questions on the personality test imagine that you are a job applicant. Please answer as you would if you were really taking the test to get a job. Keep in mind that your answers will be kept completely anonymous. 57 After these instructions were given (for the applicant condition), half of the participants were given the following instructions: To make this situation more like an applicant situation, we are offering an incentive. Those of you who score in the top 15% on this test will receive $20. Also prior to taking the test in either condition (depending on which administration was given first, and after the general instructions (honest or applicant) were given), half of the participants were told that a lie scale was included in the measure. These individuals were told the following: Sometimes people try to make themselves out to be better than they really are. Therefore, this test includes a scale to check on this. Such scales are used to identify individuals who give inaccurate or false responses. Individuals who have a high score on this scale (i.e., those individuals that the lie scale identifies as faking responses) will be eliminated from the selection process (and from consideration for the incentive if they are in that condition). Please keep this fact in mind as you take this selection test. Therefore, it was made clear to those that were in the incentive and warning condition that those who were caught faking would not be eligible for the money. After the necessary instructions were given, the personality test was administered. After taking the test under the applicant condition, participants were given the Post- Questionnaire that contained the items for the final four measures (knowledge of construct being measured, intention to fake, valence toward doing well on the test, and concern with being identified as a faker). This was followed by the administration of the demographic form. After the experiment was complete, participants were given a debriefing form (see Appendix F for a reproduction of this form) and were debriefed. 58 Results Descriptive Analyses The equivalence of the eight between-subject conditions was examined to ensure that differences between these groups were a result of the manipulations. AN OVAs and chi-squares were conducted to ensure that the eight groups were equivalent in terms of age, race, sex, GPA, past experience with personality tests, and all pre—questionnaire scales (e.g., pre-intentions to fake). When controlling for the number of analyses that were conducted (e. g., controlling for Type I error with a Bonferroni correction), no significant differences were found, indicating that the eight groups did not differ on these variables. Table 3 presents the means and standard deviations for each of the measures across the two conditions and the correlations among all the measures. There were moderate intercorrelations between the different personality scales. However, the intercorrelations of the personality scales in the honest condition are similar to normative information provided by the test publishers (Costa & McCrea, 1989). 59 on. em. 5. g. 3. 3. an... 3.. na: mu... mm: Nu. mo. 8... EN Om; EeEowmcmfi 5685:: .3 E. on. we. mN. an. on: mu: 5.- we; 3.. co. 5... No... bad :-._ scavenge-howd— ow. cm. 00. cm. 5... ha: 5.. mm: mu: we. mg. 8. mos mnAm mmocmsocc2om=ou.m_ R. ma. an. mm: .. mm: a: ma: me. No. co. :6 SNm Ease—amoch/x .2 on. mw. cm: co. we. we. vo. 3. 8. mo. mmd vmdm 80:25.9 vw. an: 3.. mo.. 3.. 8.- 5. oo. ac: Sac wmdm commao>obxmd mm. :- Nc. mo. 2. mo.- 8... war mm.w Eda Em_o:oSoZ.w ZOHEDZOU HmmZOE am. am. hm. Nb. 5. we. Nor mu. Sum exam— 8 22635.05 .N. 8. 8. mm. 8. 8.- 8.- 8. 3% 388.8255 “52089—6 8. mm. 8. 8.- 8. 8. com 2:02 03.8358 8. 8. 8. 8.- 8. mg 83.8 Emacs 83:2. .4 mas/42299320 -5: S.- 2. 8. as 2.8. 5:285 .33 cocotomxm .m E.- S. .3 Ema 88 8.2 a... ._ muiaéeozmo 3 2 a Z 2 o w h o m v m N _ .Q.m :32 8:532 Ea 853:; 2:92.onch w=o8< Scum—BSD .m DBMH 60 5.- mo. 3. 5. 3. mo. 5.. 55. No. 55. co. 5.- 5.- 55. mod on; “560?:va 22829:— .mm 5. 5.- 5. no 3. no. 5.- 5.- 5. m5. m5. 5.- co. 5.- wwd 5N4 535805-5635 3.. 3.. pm: 3.- 3. 3.- NN. cm. 3. 3. an. 3.- 5.- 5.- :6 mod mmocmsoucofimconvbm 3.- 5. 3.. me: no. 3.. 3. 3. 3. 3. 3. 5.. co. 5.- 5% 5d mmocoEmonobxm.mm mo. 5. mm. 3. 5.- 3. mm: 5.. 5.- 5.- 5.- we. no. 5. ems 5w- EflosquZNN mum—00m MUmemmn—HD 3. R. 3. ca. 3. 3. 5a.. 3.- 3.- 5.- 3.- 5.- 5.- 5.- 5:» mod Eva-Homage: cognac: . _ N R. 5. ea. 5. 3. 3. 3.. 3.- 5. m5.- 5.- 55.. 5.- 5.- 556 SN cosmooovéumdm 3. 3. 3. 3. 3. ca. 3.- 5.. 55. co. 5.- 5.- 3. co. mud Exam 805525562006— 3. 3. an. 55. m3. 5N. ma: 3.- 5.- 5.. 3.- 5. 5. 55. Now 5.3 amuse—naoohw<.w_ 3. cm. 55. 5. 3.. 3. 3.- 5. 5. mo. 5. 3. 5. 5. 53.0 3.3 86530.2 3. 3. 3N. 3. 5N. Nb. ma: 5.- no. 5.- 5. oo. 5. 5.- who when 58335372 .. mm: ma: mm: but mm: 55. 3. 5.- No. 5. mo. 5. 5.- 8.5 5.2 Em_o:oSoZ.m_ ZOEHQZOU HZobxm .0— 3. 523232 .2 zoEmzou 555.54 mm mm ow mm «m mm mm _N om o_ w. h. o. m. .3558 8:5on 98 mofimta> oEmeoEon—wdofi< macaw—ebony .m 035. 62 .3. .v- m 32 Eon E muons—280 .802 2.. mo: 3. we. 3.- 3.. 3.- 8.- 8.- 3.- 3.- MN: 3.- 5.. 3.. 3.- 3.. 3. 3.- 23m 05 NE @88qu wEom 5;? 58:00 .mm ow. 3N. we. 5. 3. 3. mo. 8.. co. 3.- 3. mm. on. ma. 3. ea. an: au. commomm< 3023.50 .8 032305— 4m om. an. an. 3. 3. 3. 5. 3. 3.. an. 3. mm. mu. .8. a. mu: No. “mo-H 5:03 mEoQ 6338. 85:3 .3 5. en. mm. m. mu. co. an. em: 3N. MN. em. 3. 3. mu. S...- 3: exam 2 32225.38 .3 ma dm 3.. 3.- NN: ac. 8.- mm. 3m. NN. 3. 3. 3.. 8. co.- 3. mmd exam 822225-5033 mao:xm V0.5 2 .om 0V0 dem 3.0 oz: 0N5 55.x. 3.x 00. _ N V0.5 0. . _ m _m.w 0:: Vm.w V0. _ N 8:808:02 Hmmzom m5. 05.5 05. V5.N N0. 05d 50. 00d o5. 05.m 05. SN 05. m5.~ 05. 50.~ 83:25-95 Va. 305 55 _w.~ 00. 05.5 mm. 5.5 mm. 0o..~ 0a. 005 M3. co.~ 5w. m5.m 0mm 0m. and V9. oV.~ m5. 00.5 N5. 2 .N mm. Sum w5. cmd 05. 5~.N mm. mad 2:82 9,0835 V5. 3:5 00. 0~.~ 00. 55.5 _0. mod 50. E .m 00. mmd N5. 0 _ .N o0. N _ .N 9.03.0 0830.5 0050? 800-9: Om :82 am :82 Om :82 00 :82 Om :82 Om :82 am :82 00 :82 80:85 0:: w:_:._:>> 80:85 80:85 0:: mEEES 0:: 3:0 ow0o_30:¥ 0:: w:_:::>> $02255— ow0o_Bo:M 3:0 30:85 3:0 305:3 802395— 880 30:00 0.80 8:000:50 80.05-:838m E _m 20 m88< 820380 080530 0:: 882 .0 20:5 69 f—vtv—.—-— Avv U» lr—r‘—~——._-p~. ‘ h....—r.-—.J o.a..j-1‘. - .-—.~A.q.|._ : .hJ. . mm. ~m.~ 2.. mm.~ 2: Sam on. Swm co.— 9: mp. w~.~ cw. mod 8.. vwd 22225-35 an. vmd wm. 2 .m we. 3d 3. E .m we. Ed _m. m3” 3.. Ed on. mud EoocoU xv. mod 9.. 86 on. 26 wv. 34m ow. E .v 3. mod mm. 3 .v 3. mod ego—39$ on. 99m K. mm.m mm. mud on. ova mm. flan :u mm.m S. m~.m 3w. om.m 85:3 820-52 $6 3 .m 36 36 2k. :6 one Sam was now an wm.m $6 ooé Ema 3.x awe: -maoscofimcou omw mm; Qum no; wad and 9: we; :1» 3; _m.v wm.m owé ma.m wed mofi mmocofiaoew< vmé 3 .- _m.m mm: own 2.. EN vm. Raw 5: 36 mm. 31‘ _c. cad mm. wmoccomo mad mo; «Ga 3. mod 3N ave mm; o _ .m KM cod mm: mm.m aqm cmh wed 56.9055 «fin SN- wwé w _ .m- mw.w Eye. 86 ohm- 5.x 3.x- 2 .m em.m mod adm- 5.x 3.x- 832.832 may—00m muzmmmmma 2:28:— 28 mafia? 03:82: «52:85 98 wEEmB 23 3:0 032305— 98 mica? 035.322 once—Bocx 2:0 95:85 3.5 9:555 among—326— 9.80 3950 28m . 32550 252280 “cubsméooiom E _m 05 383‘ mcocflgn 28.85 was 9802 .m 03$. 70 dozflgn 2.85% 8.03 2: .3 @636 was 22:38 anzmma 2: E :38 2: Bot c2228 :25: 05 E :88 05 wczoflfizm 3 622228 22> moat. Seam .2an ESQ cm a so @2885 Son 203 EoEowmcaE 5685:: can cosmoooqhom an: 35 co 8.58 kiwi $586,: $2? :wE 5:5 038 tog: ESQ 3c a 5.3 383mm =w 203 35308 3:28th 05. .me n m mug—mam =m 8m .202 EoEommENE wv. So. wed- nod 3N n _ .m 91 568:3: 3‘. So. :6- Ed mad mm.m :4 Samoan—£00m 3. So. mic fl - No.0 Rdm _ _s an. _ m 3052222850 ow. So. Ed _ - Rum Ream 2 .o modm amuse—smoew/x <2 .3 a. _- 3m 3.8 mg 8.8 32:25 Vm. _oo. 51:- $6 afivm who omdm co§o>obxm me. :5. 3.2 $5 3.2 2.x _ _dm EflouoSoZ u m 02?; mm. :82 dd :32 29.50200 HZ