A First Step in Skepticism Research:

advertisement
DEVELOPMENT OF AN INSTRUMENT TO MEASURE PROFESSIONAL
SKEPTICISM
R. Kathy Hurtt
University of Wisconsin – Madison
Grainger Hall
975 University Avenue
Madison, WI 53706-1323
(608) 265-6083
khurtt@bus.wisc.edu
November 2001
Development of an instrument to measure professional skepticism
ABSTRACT
This paper presents the development of an instrument designed to measure professional
skepticism. Professional skepticism, while acknowledged in the professional literature as being a
fundamental requirement of an audit, has received only limited academic research, and this
research has often called for development of a professional skepticism scale. Theoretically
grounded in philosophy with additional insights from professional accountants, a 30-item scale is
presented that has been tested on over 200 student and 200 professional subjects. The scale
demonstrates inter-item reliability and test-retest stability.
KEYWORDS: professional skepticism, skepticism, auditor skepticism, scale development
1
Development of an Instrument to Measure Professional Skepticism
INTRODUCTION
This paper presents the development of an instrument designed to measure professional
skepticism. Professional skepticism is a fundamental precept for the auditing profession, and the
first Statement of Auditing Standards (SAS) stressed this concept. Professional skepticism was
re-emphasized in the 1997 revisions to SAS No.1, which states:
Due professional care requires the auditor to exercise professional skepticism.
Professional skepticism is an attitude that includes a questioning mind and a
critical assessment of audit evidence. The auditor uses the knowledge, skill, and
ability called for by the profession of public accounting to diligently perform, in
good faith and with integrity, the gathering and objective evaluation of evidence.
Gathering and objectively evaluating audit evidence requires the auditor to
consider the competency and sufficiency of the evidence. Since evidence is
gathered and evaluated throughout the audit, professional skepticism should be
exercised throughout the audit process. The auditor neither assumes that
management is dishonest nor assumes unquestioned honesty. In exercising
professional skepticism, the auditor should not be satisfied with less than
persuasive evidence because of a belief that management is honest. (SAS No. 1,
AU sec. 230, paragraphs 7 through 9 (AICPA) 1997)) (highlighting added).
This emphasis in the standards suggests that research into professional skepticism is appropriate,
but it has been difficult to investigate this characteristic and its impact on the audit process.
When investigating individual differences, accounting researchers often measure the
construct of interest using scales developed in other disciplines. (A few recent examples of this
type of research include: (Bierstaker and Wright 2001; Hyatt and Prawitt 2001; Kaplan and
Whitecotton 2001; Wright and Davidson 2000).) However, professional skepticism is not a
construct that is significantly used in other disciplines. The accounting profession is unique in its
2
mandate of professional skepticism1, and consequently, no other discipline has developed an
instrument to measure this2. Even in accounting research, researchers have defined and measured
skepticism distinctively in the few accounting studies that attempt to measure this construct. A
brief summary of the extant accounting literature, the definition of skepticism used in each study
and the method used to measure the skepticism construct follows.
McMillan and White (1993) indicate, “...professional skepticism implies that auditors
focus more on error-related evidence.” They gauged professional skepticism by categorizing
subjects’ initial explanations for a hypothetical company’s financial ratio fluctuations as either an
environmental (i.e., naturally occurring) or error-based explanation. Those subjects who used an
error-based explanation were judged to be more skeptical. Shaub (1996) characterized
skepticism as the opposite of trust and operationalized the construct using the trust and
independence portions of Wrightsman’s Trustworthiness Scale (1974). Shaub and Lawrence
(1996) defined skepticism as the behavior of confronting a client or engaging in expanded audit
work and operationalized this construct through their subjects’ responses to nine scenarios. They
used a similar definition and method of measuring skepticism in their 1999 work. Choo and Tan
(2000) measured professional skepticism using ten items from the Rempel, Holmes and Zanna
(1985) Trust Scale. Choo and Tan modified the scale so that the questions referred to a “client”
rather than the life-partner referenced in the original scale. Subjects with low scores on the Trust
Scale were assumed to have high skepticism. Choo and Tan conclude their study with a
1
Other professions such as attorneys, police officers, and bank examiners can benefit from skepticism; however, no
other discipline has a requirement for professional skepticism or the emphasis on professional skepticism that is
present in accounting.
2
There have been a few instances in marketing where researchers investigated skepticism, but these scales are so
specifically directed towards skepticism about advertising that they are of no assistance in measuring professional
skepticism.
3
discussion of the problems of using the Trust Scale and call for the development of a
psychometrically reliable scale designed specifically to measure skepticism.
Using unique measures of skepticism within each study makes it difficult to develop an
understanding of the behavioral implications of professional skepticism. In some instances,
because of the embedded nature of the scales in the experiment, the scale used cannot easily be
replicated in a subsequent study3. In other instances, the amount of time required for subjects to
complete the skepticism measure is so great, that it would be difficult to use the measure in most
experimental contexts4. In other instances, the authors described scales they used in their study
as easily available proxies that were used only because an instrument designed specifically to
measure professional skepticism had not been developed (Shaub, 1996, Choo and Tan, 2000).
This paper presents the development of such an instrument: one designed specifically to measure
professional skepticism. In the remainder of the paper I will present a brief overview of the steps
required for instrument development, describe how those steps were completed for the
professional skepticism instrument developed here, and discuss future research directions.
INSTRUMENT DEVELOPMENT
Scale development must flow from theory (DeVellis, 1991). Hurtt, Eining and Plumlee
(2001) present a theoretical model of professional skepticism that is grounded in psychological
literature and philosophical literature on methodological skepticism. Their model (presented in
Figure 1) suggests professional skepticism is a multi-dimensional construct comprised of six
individual characteristics: curiosity; self-confidence; interpersonal understanding, questioning;
self-determining and deliberating.
3
In private conversation, Jeffrey McMillan indicated that the embedded nature of their instrument would make
replicating their skepticism measure for comparison to other measures almost impossible.
4
In private correspondence with Shaub and Lawrence, they estimated that it took subjects approximately 25 minutes
to complete the scenario section of their instrument.
4
Insert Figure 1 about here
They describe their curiosity characteristic as searching for knowledge and investigating.
Hurtt et al. (2001) differentiate curiosity from questioning by characterizing questioning as a
requirement for reasons, evidence, justification or proof, rather than the personal investigation
that characterizes curiosity. Self-confidence, while self-explanatory, is included in their model
because of the necessity of a skeptical individual to have the courage to act on their curiosity and
questioning. Interpersonal understanding is defined as a curiosity about people that is more
specific than the general trait of curiosity. Self-determining is characterized as a reluctance to
accept others’ statements or claims without making a self-determination, and deliberating is the
characteristic of reflective decision-making. Using these elements, I followed standard scale
development techniques to create a scale designed specifically to measure professional
skepticism.
Developing an initial item pool
Clark and Watson (1998) suggest that after a theoretical basis for the scale is identified,
existing literature should be reviewed to determine how others have assessed the same or closely
related constructs. I reviewed the psychological scales suggested by Hurtt, et al. for items
designed to measure constructs closely related to one or more of the six elements of skepticism,
and I also performed an independent search for any existing scales designed to measure
professional skepticism or a similar construct. Although I did not find an existing scale, from the
scales examined I identified over 170 individual items that could be modified to capture one of
the six skepticism elements. Table 1 indicates the scales from which items were extracted for the
initial item pool5. Some elements of professional skepticism did not correspond well to items
5
In many instance, I theorized that individual items from an existing scale would likely map to more than one
construct in the professional skepticism scale. This splitting of items from existing scales is due to sub-constructs
5
from existing scales and to resolve such instances I wrote an additional 50 items. This created an
initial item pool of 220 questions. Two Ph.D. students with experience teaching scale
development and I reviewed each of the items for clarity and to eliminate redundant questions.
This process reduced the number of items to 85.
Insert Table 1
To allow for consistent presentation, both those items that had been used in previous
scales and all the new items were written for use with a 6-point6 Likert type scale anchored by
“strongly agree” and “strongly disagree.” Changing item wording, scaling methods or anchors
invalidates established reliabilities for measures (Converse and Presser 1990, Schwarz 1996);
therefore, all the items used in this new instrument needed to be re-validated, regardless of the
reliability or validity of the original scale.
One of the major threats to construct validity is construct under-representation. This
occurs when assessment of a construct is too narrow and fails to include important dimensions of
the construct (Messick 1998). Because the Hurtt et al. (2001) model was developed from
psychology research and from philosophical skepticism, I wanted to see if this theoretical view
of professional skepticism differed from that of practicing accountants. I solicited written
comments on skepticism from thirty-five professionals (corporate accountants and auditors) and
asked them to answer open-ended questions about what they believed comprised skepticism and
within the original scale. For example, items in the Melbourne Curiosity Inventory (MCI), (Bunge 1991), that dealt
with general curiosity were initially included as curiosity questions in the professional skepticism scale, while those
items that dealt with curiosity about people were included as interpersonal understanding questions. This results in
the same scale being listed under different constructs in Table 1.
6
There is considerable discussion in the scale development literature about whether an odd or even number of
response points is preferable (see for example, Moser and Kalton, 1972, 344[Moser, 1972 #215] and Rugg and
Cantril, 1944, 33). With an odd number of response alternatives, the subjects are offered a neutral middle. I used an
even number to avoid the issue of subjects simply selecting the mean value. This is consistent with the
recommendation of Converse and Presser (1990, 37)[Converse, 1990 #209] who indicate, “Do not explicitly
provide the middle category and thereby avoid losing information about the direction in which some people lean….”
6
what they believed would be the opposite of skepticism. I compared these comments to both the
theorized constructs in the model and to the individual scale items. This review did not identify
any dimensions of professional skepticism that were not included in the model, and the
definitions of skepticism were consistent with the items used in the scale.
DeVellis (1991) also recommends that experts review the scale. This process is intended
to improve the instrument’s content validity. I met with three accounting faculty members with
backgrounds and education in auditing who reviewed the scale items for relevance and to ensure
that they agreed that each item was properly classified regarding the construct the item was
designed to measure. I eliminated items identified by the accounting faculty members as
irrelevant or inconsistent.
Pre-testing the instrument
From this process I reduced the scale to 49 items and administered the instrument to 89
graduate and undergraduate business students in what Converse and Presser (1990) describe as a
“developmental pretest.” The intent of this pretest is to smooth the wording of the questions and
to see which questions measure the constructs under consideration. They recommend that this
first pretest contain less than twice as many questions as the final instrument will contain, that
the respondents resemble the target population and that a reasonable number of subjects for this
stage pretest is 25 – 75. Because the instrument is designed to be used both with business
students, and also with professional accountants, the use of a business student subject pool for
this pretest is justified.
I examined the results of the pretest using factor analysis to assist me in identifying items
that did not load on expected constructs. My initial examination of the correlation matrix
They further suggest measuring intensity by anchoring with words such as “strongly agree” and “strongly disagree”
as was done in this scale.
7
indicated that many items were significantly correlated; therefore, I used oblique rotation (direct
oblimin), which relaxes the assumption of orthogonal factors. After eliminating items that did
not load on any of the hypothesized factors, 22 items remained. These 22-items loaded on the
six theoretical factors with item factor-loadings generally of .70 or higher (Table 2). Factor
loadings above .50 are considered acceptable, (DeVellis, 1991).
Insert Table 2 about here
Cronbach’s alpha, which indicates whether the questions in the scale are reliably
measuring the same construct, for the entire scale was .827. Cronbach’s alpha was also
calculated for each of the sub-scales. For five of the six dimensions of skepticism, the alpha for
the sub-scale exceeded .80. Alphas greater than .80 are considered adequate results, although the
small sample size and fairly homogeneous composition of the subject pool could lead to
overestimating the reliability of the items. Although the sub constructs are not to be used
individually, and the alpha for the entire scale was adequate, the alpha for the “questioning”
construct was very low at .57. To address this problem, I generated a number of new items for
this construct that could be used in a subsequent pilot test. I also closely examined each item by
construct to identify any items that had been confusing to the subjects, and I wrote or revised
additional items towards the goal of obtaining an equal number of items (5) for each of the six
factors. Because there is no theoretical basis for assuming that any single aspect of skepticism is
more important than another, developing a scale with equal weighting for all factors seemed
DeVellis (1991) indicates the following acceptable ranges for Cronbach’s alpha: “between .80 and .90 is very
good; between .70 and .80, respectable; between .65 and .70 minimally acceptable.”
7
8
most appropriate8. This process resulted in a 40-item measure for the pilot test. The pilot test had
more items than I wanted for the final test to allow for the possibility of invalid questions.
Pilot and reliability testing
The revised 40-item scale was administered to 250 undergraduate business students and
useable responses were received from 247 students. Seventeen responses had at least one item
skipped. The mean score for the item was substituted in these instances so that the remainder of
the response could be analyzed9. Analysis of the data revealed no significant differences in
skepticism scores related to any of the demographic information such as gender, GPA, class
standing, class where the instrument was administered, native English speaker, or age.
The correlation matrix of the responses indicated significant or marginally significant
correlations for many of the items; therefore, I again used oblique rotation for the factor analysis
of the responses. The rotated-factor analysis indicated that the majority of items loaded on the
six theoretical factors. I eliminated items that loaded on constructs other than those anticipated
(this occurred primarily with new or reversed items), which resulted in a 25-item scale. Table 3
lists the factor loadings for these items.
Insert Table 3 about here
Cronbach’s alpha for the entire scale was .85, but, as expected given the small sample
used for the previous test, both factor loadings and Cronbach’s alpha decreased for most of the
sub-constructs. These results indicate that overall the professional skepticism scale has adequate
inter-item reliability, although the sub-construct “questioning” remained below the normally
8
30-items is a parsimonious number for a scale containing 6 sub-constructs; however, my goal was to develop an
instrument that could be quickly administered before or after behavioral experiments, thus a fairly brief scale was
required.
9
accepted range. Although not ideal, the low alpha for this sub-construct indicates that these
questions are not all measuring the same characteristic. Because the scale as a whole
demonstrates adequate reliability, the overall score should be a reliable measure. However, I
again wrote additional questions to attempt to raise the alpha score for the questioning construct
and to ensure that there were an equal number of questions for each construct.
Intra-subject test-retest reliability measures the stability of an individual’s responses over
time. The test-retest method is recommended when developing a measure that cannot be
validated using known group10 validation techniques, or when developing a measure with subconstructs that preclude using split-half reliability (DeVellis, 1991, Fink, 1995, Litwin, 1995).
After an approximately 8-week delay, the 250 subjects were asked to complete the professional
skepticism instrument for a second time11. The subjects were also asked to complete
Wrightsman’s (1974) Trustworthiness and Independence Scales that were used by Shaub
(1996)12. Having subject’s scores on the professional skepticism instrument as well as the two
Wrightsman’s scales allowed me to begin the process of determining whether the skepticism
scale measures a different construct than the Wrightsman scales, (i.e., discriminant validity).
Responses were received from 147 subjects, but because of incomplete information only
92 matched sets of tests were identified. The correlation between the total skepticism score (sum
of the 25 items) on the first and second test was significant, r (92) = .89, p<. 001. Litwin (1995)
indicates that in reliability testing, “r values are considered good if they equal or exceed .70,”
9
Although the factor loadings changed, there were no significant differences if the median was substituted or if
these questions were dropped.
10
Known group validation involves administering the instrument to a group of subjects known to have the
characteristic that the scale is designed to measure. Because there is not a “known group” that is skeptical, this
preferred validation technique cannot be used.
11
Although questions were added to bring the total number of scale questions again to 40, these additional questions
were not included in the test-retest analysis. Only the twenty-five questions displayed in Table 3 were included in
the analysis.
12
Shaub found that these scales were not predictive of the skeptical behaviors that he expected from his subjects.
10
while Fink (1995) suggests “a conservative rule of thumb: over .75 is a very good to excellent
relationship.” Cronbach’s alpha for the combined test-retest was .95. The 15 “new” items were
examined to identify questions that loaded on the appropriate factor. From this a 30-item test
was constructed with an equal number of items for each construct. (See appendix A for the 30item Professional Skepticism Instrument.) On this 30-item instrument the subjects had a mean
score of 132.7 with a standard deviation of 15.9, and scores ranged from 77 to 175.
As shown in Table 4, subjects’ score on the professional skepticism scale was not
significantly correlated with either Wrightsman’s Trustworthiness or Wrightsman’s
Independence Scales. This indicates that the professional skepticism scale is measuring a
different construct than either trustworthiness or independence, and might help to explain why
Shaub (1996) found neither test was predictive of his subjects’ decision to exercise skeptical
behavior. This presents preliminary evidence that a scale designed specifically to measure
professional skepticism can assist in further study of this important construct.
Insert Table 4 about here
Testing with professional accountants
I received permission from a major international accounting firm to administer the
professional skepticism instrument to members of their firm. Managing partners at thirteen U.S.
offices of this firm agreed to provide email addresses for their staff members13 and also agreed to
send a memo to those individuals encouraging them to respond to my email. I emailed 350
auditors and asked them to participate in a two-part study whose purpose was represented as
being research on the work paper review process. The email included the URL for a web site that
13
The partners were asked to supply email addresses for individuals with four or more years of audit experience.
This was requested as part of an experimental study that was also conducted at this time; however, it also served to
ensure that only those individuals who have experience with auditing and the requirement that auditors exhibit
professional skepticism were included in this instrument validation.
11
contained the professional skepticism instrument, and questions on demographic information. To
increase the response rate, I sent two follow-up emails during the next two months. Two hundred
auditors completed the questions at the designated URL, a 57% response rate. Tests of
differences between early and late responders did not reveal any differences between the groups.
Demographics on the auditors are presented in Table 5. Their mean score on the professional
skepticism scale was 138.6 with a range from 111 to 173 (of 180 points possible), and a standard
deviation of 12.6. There was no significant difference between the subject’s score on the scale
and any of their demographics (e.g., age, amount of experience, or gender.) The mean experience
level of the auditors was 58 months, so all of the subjects should be familiar with the requirement
for professional skepticism. However, the purpose of the scale was not disclosed to the subjects
to lessen the likelihood of subjects responding in a manner that they believed was “correct.”
Insert Table 5 about here
The internal consistency coefficient for this administration of the 30-item scale using
Cronbach’s alpha is .86. As with previous administrations of the instrument, and as expected for
a scale designed to measure a single construct, the correlation matrix indicated that many of the
items were significantly correlated; therefore, oblique rotation was used in the factor analysis of
the responses. As shown in Table 6, the questions loaded on the six theorized factors, but two
questions did not load on the construct they were theorized to measure, and one of these items
loaded at only .49, below the .50 minimum suggested by DeVellis (1991)14. Because this scale is
designed to provide a single score for professional skepticism, and the internal consistency of the
14
It is unknown if this is a problem unique to this sample population, or a problem with the instrument. Only
further testing with a broad variety of subjects will be able to answer that question. However, I did rerun the
analysis using the combined first and second scores of the subjects and all items loaded on the appropriate
constructs.
12
questions is appropriate, this shift in where the two questions loaded might be simply a function
of the sample population.
After a brief wait, subjects were again emailed and asked to complete an online work
paper review, answer a few questions and then complete the skepticism scale. Scores on the
skepticism scale were collected from 88 subjects. Their mean score on the second test was 135.4,
with a range of 105 to 177 and a standard deviation of 14.7. This contrasts with these 88 subjects
scores on the first administration of the professional skepticism scale of a mean of 138.5, a range
of 113 to 170 and a standard deviation of 12.4. The mean elapsed time between completion of
the first and second scale was 22 days, with a range from 2 days to 56 days, and a mode of 15
days. The correlation between the first and second test score was .85, which is significant at the
.01 level. Cronbach’s alpha, measuring internal consistency between the two total scores was
.91.
SUMMARY DISCUSSION
This paper presents the development of an instrument designed to measure professional
skepticism. Previous researchers have used proxy measures or experimental type measures that
are lengthy to administer and time-consuming to score. The scale presented here is an initial
effort at developing a psychometrically valid instrument that can be easily administered and
quickly scored by researchers interested in the area of professional skepticism. A limitation of
this development process is that the instrument was validated using business students at a single
school and professionals from a single firm. Research using other psychological-based
instruments has demonstrated that there can be significant differences between the average
scores between firms, and that possibility should be explored in future research.
13
The theoretical foundation for the professional skepticism instrument was an examination
of philosophical writings about skepticism for elements that are descriptive of professional
skepticism. In an effort to identify any elements of professional skepticism that were not evident
in philosophical skepticism, accounting research was reviewed and input was solicited from
accounting professionals. These efforts did not identify any characteristics that were not present
in the philosophical model.
The analysis of the instrument indicates that the scale has adequate inter-item
consistency, and adequate test-retest reliability with both student and professional subjects. This
is a necessary first step in instrument development; however, significantly more research needs
to be completed. The validity of the instrument, which is whether it actually measures the
professional skepticism that it purports to measure, must be determined. The model in Figure 1
predicts that the characteristics of skepticism should lead to four behaviors. Research needs to be
undertaken to determine if this instrument differentiates between subjects and whether different
scores lead to these different behaviors. Additional issues such as the longitudinal stability of the
measure over a professional career and whether a subject’s score changes through education and
coursework also should be examined.
Research on professional skepticism has been difficult because of the lack of a standard,
easily administered method of determining differences in this characteristic between subjects.
With the availability of an instrument, a host of interesting research questions is possible such as
“Does a higher level of professional skepticism lead to increased fraud detection?” “Can an
auditor be too skeptical?” “Are more skeptical auditors slower at performing audit work: does
professional skepticism lead to over-auditing and inefficient audits?” “Are there differences in
the review procedures of more and less skeptical auditors?” “Can professional skepticism be
14
taught or learned?” “Is professional skepticism related to expertise, or schemas, or some other
type of cognitive process?” The development of a psychometrically valid instrument to measure
professional skepticism is perhaps overdue, and will hopefully spur additional research into an
area that is both professionally mandated and is often cited as lacking in SEC actions against
auditors.
15
Figure 1
Model of Professional Skepticism
(Hurtt, Eining and Plumlee, 2001)
Behaviors of Skeptics
Characteristics of Skeptics
Expanded
Information
Search
Curiosity
Self-confidence
Interpersonal
understanding
Questioning
Individual’s
Level of
Professional
Skepticism
Increased
Contradiction
Detection
Increased
Alternative
Generation
Self Determining
Increased Scrutiny
of Interpersonal
Information
Deliberating
The model of professional skepticism addresses both characteristics and behaviors. The
professional skepticism instrument only measures the individual characteristics. These
characteristics are in the dotted portion of the model.
16
Table 1
Characteristics of Skeptics
(Adapted from Hurtt, Eining and Plumlee, 2001)
Characteristics of
Skeptics
Related Philosophical
and Psychological
Literature
Psychological Scales Examined
for initial item pool
Curiosity
Bunge (1991)
Johnson (1978)
Naylor (1981)
Spielberger (1995)
Melbourne Curiosity Inventory (MCI)
(Naylor 1981)
State-Trait Personality Inventory (STPI)
(Spielberger 1995)
Self-confidence
Forsman and Johnson (1996)
Fleming and Courtney (1984)
Helmreich and Stapp (1974)
Linn et al. (1982)
Interpersonal
understanding
Questioning
Self-determining
Deliberating
Burnyeat (1983)
Hallie (1985)
Hookway (1990)
Johnson (1978)
Kurtz (1992)
McGinn (1989)
Popkin (1979)
Dimensions of Self Esteem
(Forsman and Johnson 1996)
Dimensions of Self Esteem II
(Fleming and Courtney 1984)
Texas Social Behavior Inventory
(Helmreich and Stapp 1974)
Melbourne Curiosity Inventory (MCI)
(Naylor 1981)
State-Trait Personality Inventory (STPI)
(Spielberger 1995)
Fogelin (1994)
Kurtz (1992)
Rempel et al. (1985)
Rotter (1967)
Rotter (1980)
Spielberger (1995)
Stough (1969)
Wrightsman (1974)
State-Trait Personality Inventory (STPI)
Spielberger, 1995
10 Item Trust Scale (Rempel et al. 1985)
Trustworthiness and Independence Scale
(Wrightsman 1974)
Interpersonal Trust Scale (Rotter 1967)
Interpersonal Trust, Trustworthiness and
Gullibility Scale (Rotter 1980)
Carver (1997)
Lefcourt (1982)
Lefcourt (1991)
Stickland (1989)
Rotter (1966)
Locus of Control (Rotter 1966)
Bunge (1991)
Hallie (1985)
Kruglanski (1990)
Kruglanski et al. (1993)
Kurtz (1992)
MacDonald (1970)
McGinn (1989)
Mikulincer (1997)
Webster & Kruglanski (1994)
Tolerance of Ambiguity
(MacDonald 1970)
Need for Cognitive Closure
(Kruglanski 1990)
17
Table 2
Rotated Component Matrix for Pretest - 23-items
1
2
Factor
3
4
Scale  = .82; n=89
5
6
.88
.87
-.22
.06
.15
.23
.15
.24
-.12
-.29
.25
.21
.85
.84
-.00
-.09
.17
.26
.26
.16
-.29
-.35
.29
.20
.67
.05
.14
-.01
-.34
.34
-.00
-.03
.03
-.06
-.19
.89
.85
.80
.79
.76
.16
.09
.06
.12
-.05
-.23
-.09
-.09
-.07
-.10
-.16
-.02
.13
-.15
-.02
.09
.03
-.11
-.21
.03
.12
.00
.89
.22
-.31
.21
.12
.21
.09
.08
.82
.78
.06
.21
-.44
-.18
.18
.22
.28
.13
.74
.25
-.18
.33
.19
.17
.17
.30
.33
.41
-.06
-.08
-.23
.04
.08
-.06
.25
.22
.20
.35
.32
.36
.94
.94
.91
.20
.22
.33
-.24
-.25
-.11
-.92
-.90
-.79
Item
Interpersonal Understanding
( = .88)
I like to understand the reason for other people’s behavior.
I am interested in what causes people to behave the way that
they do.
I like thinking about how individual people influence events.
The actions people take and the reasons for those actions are
fascinating.
I often find myself imagining scenarios to explain why people
behave in a certain way.
Deliberating
( = .87)
I seldom notice inconsistencies in explanations.
Most often, I agree with what the others in my group think.
I tend to immediately accept what other people tell me.
I usually accept things I see, read, or hear at face value.
Bad luck or good luck can really follow you around.
Self-determining
( = .82)
I like to ensure that I’ve considered most available
information before making a decision.
I take my time when making decisions.
I don’t like to decide until I’ve looked at all of the readily
available information.
I carefully consider most information before I accept it as
true.
.11
.07
.13
Self-confidence
( = .93)
I have confidence in myself.
I feel sure of myself.
I feel good about me.
.17
.14
.16
Curiosity
( = .87)
I think that learning is exciting.
The prospect of learning excites me.
Discovering new information is fun.
.17
-.05
.23
.08
-.13
.79
.24
.32
.04
-.03
.43
.12
.14
.10
-.31
.01
.71
.69
Questioning
( = .57)
My friends tell me that I often question things that I see or
hear.
I enjoy trying to determine if what I read or hear is true.
I often doubt other people’s explanations.
18
Table 3
Rotated Component Matrix for the Pilot - 25-items
1
2
Factor
3
4
Scale  .85; n = 247
5
6
Item
(“R” = Reversed item)
.90
.90
.81
.76
.16
.17
.15
.41
.10
.13
.11
.02
-.22
-.30
-.44
-.30
.10
.07
.12
.27
.35
.44
.48
.38
.24
.01
.31
.79
.74
.74
.08
.00
.23
-.22
-.03
-.13
.21
.02
.53
.08
-.12
.11
.42
.73
.06
-.11
.33
.22
.06
.00
.06
.25
.11
.13
-.13
.00
.24
.17
.77
.73
.71
.66
.66
-.16
-.07
-.16
-.18
-.10
.08
.01
.20
.09
.32
.35
.16
.00
-.86
.08
.31
.26
.19
.22
.10
.06
.08
.22
.29
-.86
-.77
-.77
.21
.03
-.06
.02
.06
.15
.42
.46
.37
.42
.30
.09
.24
.21
.27
.04
.11
.02
-.04
.11
.11
.20
.18
.03
.14
.16
.26
-.07
-.07
.19
-.28
-.18
-.16
-.19
-.21
.69
.68
.63
.62
.12
.06
.21
-.08
Curiosity
( = .87)
I think that learning is exciting.
The prospect of learning excites me.
Discovering new information is fun.
I like learning more about many situations.
Deliberating
( = .75)
I take my time when making decisions.
I dislike having to make decisions quickly.
I don’t like to decide until I’ve looked at all of the readily
available information.
I like to ensure that I’ve considered most available information
before making a decision.
.26
.08
.12
.14
.09
Self-determining
( = .76)
It is easy for other people to convince me. (R)
I let what others say influence my decision making. (R)
Most often I agree with what the others in my group think. (R)
I usually I accept things I see, read, or hear at face value. (R)
I tend to immediately accept what other people tell me. (R)
.20
96
.22
.21
.10
Interpersonal Understanding
( = .84)
I am interested in what causes people to behave the way that they
do.
I like to understand the reason for other people’s behavior.
I seldom consider why people behave in a certain way. (R)
Other people’s behavior doesn’t interest me. (R)
.14
-.17
.04
.31
.88
.86
.86
.79
Questioning
( = .60)
My friends tell me that I often question things that I see or hear.
I usually notice inconsistencies in explanations.
I often reject statements unless I have proof that they are true.
I enjoy trying to determine if what I read or hear is true.
Self-confidence
( = .87)
I have confidence in myself.
I feel good about myself.
I am self-assured.
I don’t feel sure of myself. (R)
19
Table 4
Correlation Matrix of Skepticism Score with Wrightsman’s
Trustworthiness and Independence Scales
Skepticism
Trustworthiness
Independence
*
p < .001, n = 140
Skepticism
Trustworthiness
.115
.057
.507*
20
Table 5
Demographics on Professional Accounting Respondents
Mean
Standard Deviation
Months of professional experience
Age
Education:
Bachelor’s degree *
Advanced degree
Certifications held:
CPA
Other
Male
Female
Declined to respond
Total
n = 112
n = 87
n=1
n = 200
60
26
28.7
3.3
57
21
27.8
3.0
75
58.4
24
28.3
3.2
90
22
72
14
99
6
79
2
30
1
1
162
37
179
8
* One subject did not complete the information on education.
21
Table 6
Rotated Component Matrix for the Professional Subjects
1
2
Factor
3
4
Scale  = .86; n = 200
5
6
Item
(“R” = Reversed item)
Curiosity ( = .88.)
I think that learning is exciting.
I relish learning.
Discovering new information is fun
I like searching for knowledge.
The prospect of learning excites me.
I enjoy trying to determine if what I read or hear is true.
.91
.83
.83
.82
.77
.55
.09
.07
.07
.16
.03
.22
.15
.18
.07
.20
.17
.18
.21
.25
.20
.25
.08
.34
.26
.21
.28
.32
.28
.27
-.14
-.22
-.11
-.10
-.12
-.45
.07
.03
.82
.80
-.06
.14
.18
.22
-.01
.02
-.14
-.16
.03
.17
.77
.76
-.03
.23
.08
.12
-.22
.09
-.25
-.13
.11
.69
.01
.06
-.14
.02
Deliberating ( = .83.)
I take my time when making decisions.
I don’t like to decide until I’ve looked at all of the readily available
information.
I dislike having to make decisions quickly.
I like to ensure that I’ve considered most available information before
making a decision.
I wait to decide on issues until I can get more information.
.07
.12
.13
.12
.08
.25
.07
.19
-.03
.08
-.03
.11
.76
.76
.74
.64
.56
.49
.07
.11
.07
.10
-.03
.05
.25
.06
.22
.29
.12
.17
-.37
-.25
-.37
-.57
.05
-.11
Self-determining ( = 76.)
I tend to immediately accept what other people tell me. (R)
I usually I accept things I see, read, or hear at face value. (R)
I often accept other people’s explanations without further thought. (R)
It is easy for other people to convince me. (R)
Most often I agree with what the others in my group think. (R)
I usually notice inconsistencies in explanations.
.28
.22
.33
.19
.19
.12
.10
-.00
-.04
.90
.88
.88
.13
.06
.02
-.21
-.18
-.10
.14
.05
.06
.12
.17
-.02
.79
.75
.08
.11
-.03
.14
Interpersonal Understanding ( = 90.)
I like to understand the reason for other people’s behavior.
I am interested in what causes people to behave the way that they do.
The actions people take and the reasons for those actions are
fascinating.
I seldom consider why people behave in a certain way. (R)
Other people’s behavior doesn’t interest me. (R)
.29
.25
.34
.22
.24
-.10
-.03
-.01
-.01
-.08
.17
.25
.23
.11
.18
.02
.11
.14
.11
.04
.90
.87
.87
.84
.81
-.06
-.07
-.12
-.15
.06
Self-confidence ( = 91.)
I have confidence in myself.
I don’t feel sure of myself. (R)
I am self-assured.
I am confident of my abilities.
I feel good about myself.
.23
.32
.00
.06
.07
.31
.21
.46
.21
.11
.12
-.02
.07
.17
.01
-.85
-.78
-.60
Questioning ( = .67.)
My friends tell me that I often question things that I see or hear.
I frequently question things that I see or hear.
I often reject statements unless I have proof that they are true
22
American Institute of Certified Public Accountants (AICPA), 1997. Due care in the performance
of work, in Statement on Auditing Standards No. 1 (New York, NY).
Bierstaker J. L., S. Wright. 2001. A research note concerning practical problem-solving ability as
a predictor of performance in auditing tasks. Behavioral Research in Accounting 13 : 4959.
Bunge M. 1991. A skeptic's beliefs and disbeliefs. New Ideas in Psychology 9 (2): 131-149.
Burnyeat M. 1983. The Skeptical Tradition.(Berkeley: University of California Press).
Carver C. S. 1997. The Internal-External Scale Confounds Internal Locus of Control with
Expectancies of Positive Outcomes. Personality and Social Psychology Bulletin 23 (6):
580-585.
Choo F., K. Tan. 2000. Instruction, skepticism, and accounting students' ability to detect frauds
in auditing. The Journal of Business Education 1 (Fall): 72 - 87.
Clark L. A., D. Watson. 1998. “Constructing Validity: Basic issues in objective scale
development”, in Methodological issues and strategies in clinical research. Ed A. E.
Kazdin (Washington, DC: Washington, DC) 215-239.
Converse J. M., S. Presser. 1990. Survey Questions Handcrafting the standardized
questionnaire.(Newbury Park, CA: SAGE Publications, Inc.).
DeVellis R. F. 1991. Scale Development: Theory and Applications.(Thousand Oaks, California:
SAGE Publications).
Fink A. 1995. How to Design Surveys.(Thousand Oaks, California: SAGE Publications, Inc.).
Fleming J. S., B. E. Courtney. 1984. The dimensionality of self-esteem II: Heirarchical facet
model for revised measurement scales. Journal of Personality and Social Psychology 46 :
404-421.
Fogelin R. J. 1994. Pyrrhonian reflections on knowledge and justification.(New York: Oxford
University Press).
Forsman L., M. Johnson. 1996. Dimensionality and validity of two scales measuring different
aspects of self-esteem. Scandinavian Journal of Psychology 37 : 1-15.
Hallie P. P., 1985. Sextus Empiricus: Selections from the major writings on skepticism, man &
God, (Indianapolis) pp 238.
Helmreich R., J. Stapp. 1974. Short forms of the Texas Social Behavior Inventory (TSBI), an
objective measure of self-esteem. Bulletin of the Psychonomic Society 4 : 473-475.
23
Hookway C. 1990. Scepticism.(New York: Routledge).
Hurtt R. K., M. Eining, R. D. Plumlee. 2001. What is Professional Skepticism? Working Paper,
University of Wisconsin: .
Hyatt T. A., D. F. Prawitt. 2001. Does Congruence between audit structure and auditors' locus of
control affect job performance? The Accounting Review 76, No. 2 (April): 263-274.
Johnson O. A. 1978. Skepticism and cognitivism.(Berkeley: University of California Press, Ltd.).
Kaplan S. E., S. M. Whitecotton. 2001. An examination of Auditors' Reporting Intentions when
another auditor is offered client employment. Auditing: A journal of practice and theory
20 (1): 45-63.
Kruglanski A. W. 1990. “Motivations for judging and knowing: Implications for causal
attribution”, in The handbook of motivation and cognition: Foundation of social
behavior. Eds E. T. Higgins and R. M. Sorrention (New York: New York) 333-368.
Kruglanski A. W., D. M. Webster, A. Klem. 1993. Motivated resistance and openness to
persuasion in the presence or absence of prior information. Journal of Personality and
Social Psychology 65 (5): 864-876.
Kurtz P. 1992. The New Skepticism: inquiry and reliable knowledge.(Buffalo, New York:
Prometheus Books).
Lefcourt H. M. 1982. Locus of Control: Current Trends in Theory and Research.(Hillsdale, NJ:
Lawrence Erlbaum Associates).
Lefcourt H. M. 1991. “Locus of Control”, in Measures of personality and social psychological
attitudes. Eds J. P. Robinson, P. R. Shaver and L. S. Wrightsman (San Diego: San Diego)
413-499.
Linn M. C., T. de Benedictis, K. Delucchi. 1982. Adolescent reasoning about advertisements:
Preliminary investigations. Child Development 53 : 1599 - 1613.
Litwin M. S. 1995. How to measure survey reliability and validity.(Thousand Oaks, California:
SAGE Publications, Inc).
MacDonald A. P. J. 1970. Revised Scale for Ambiguity tolerance: reliability and validity.
Psychological Reports 26 : 791-798.
McGinn M. 1989. Sense and Certainty: A dissolution of scepticism.(New York: Basil Blackwell
Inc.).
24
McMillan J. J., R. A. White. 1993. Auditors' Belief Revisions and Evidence Search: The Effect
of Hypothesis Frame, Confirmation and Professional Skepticism. The Accounting Review
July : 443-465.
Messick S. 1998. “Validity of psychological assessment: validation of inferences from persons
responses and performances as scientific inquiry into score meaning”, in Methodological
issues and strategies in clinical research. Ed A. E. Kazdin (Washington, DC:
Washington, DC) 241-260.
Mikulincer M. 1997. Adult attachment style and information processing: individual differences
in curiosity and cognitive closure. Journal of Personality and Social Psychology 72 (5):
1217 - 1230.
Naylor F. D. 1981. A state-trait curiosity inventory. Austrailian Psychologist 16 (2): 172 - 183.
Popkin R. H. 1979. The history of scepticism from Erasmus to Spinoza.(Berkeley: University of
California Press).
Rempel J. K., J. G. Holmes, M. P. Zanna. 1985. Trust in close relationships. Journal of
Personality and Social Psychology 49 : 95-112.
Rotter J. B. 1966. Generalized Expectancies for Internal versus external control of
reinforcement. Psychological Monographs: General and Applied (Whole No. 601) 80,
no. 1.
Rotter J. B. 1967. A new scale for the measurement of interpersonal trust. Journal of Personality
35 : 651-665.
Rotter J. B. 1980. Interpersonal trust, trustworthiness, and gullibility. American Psychologist 35
(1): 1 - 7.
Schwarz N. 1996. Cognition and Communication: Judgmental Biases, Research Methods, and
the logic of conversation.(Mahwah, New Jersey: Lawrence Erlbaum Associates).
Shaub M. K. 1996. Trust and suspicion: The effects of situational and dispositional factors on
auditors' trust of clients. Behavioral Research in Accounting 8 : 154-174.
Shaub M. K., J. E. Lawrence. 1996. Ethics, experience and professional skepticism: A
situational analysis. Behavioral Research in Accounting 8 (Supplement): 124-157.
Shaub M. K., J. E. Lawrence. 1999. Differences in auditors' professional skepticism across
career levels in the firm. Advances in Accounting Behavioral Research 2 : 61 - 83.
Spielberger C. D., 1995. State-Trait Personality Inventory (STPI), (Redwood City, California) .
25
Stickland B. R. 1989. Internal-external control expectancies: From contingency to creativity.
American Psychologist 44 : 1-12.
Stough C. L. 1969. Greek Skepticism.(University of California Press, Ltd.).
Webster D. M., A. W. Kruglanski. 1994. Individual differences in need for cognitive closure.
Journal of Personality and Social Psychology 67 (6) : 1049-1062.
Wright M. E., R. Davidson. 2000. The effect of auditor attestation and tolerance for ambiguity
on commercial lending decisions. Auditing: A journal of practice and theory 19 (2).
Wrightsman L. S. 1974. Assumptions about human nature: A social-psychological
approach.(Monterey, California: Brooks/Cole Publishing Company).
26
Download