Document 10464497

advertisement
International Journal of Humanities and Social Science
Vol. 2 No. 14 [Special Issue - July 2012]
The Assessment of the Adult Learning and Development Student Satisfaction Scale
(ALDSS Scale)
Jonathan E. Messemer, Ed.D.
Catherine A. Hansman, Ed.D.
Cleveland State University
College of Education and Human Services
CASAL Department – Julka Hall 275
2121 Euclid Avenue
Cleveland, Ohio 44115-2214
United States of America
Abstract
The purpose of this article is to discuss the development and assessment of the Adult Learning and Development
Student Satisfaction Scale (ALDSS Scale). The ALDSS Scale was designed to measure student satisfaction
regarding the following six factor groups: (1) curriculum, (2) learning format, (3) course materials, (4) program
access, (5) faculty and instruction, and (6) faculty advising. During a three year period, the ALDSS Scale was
administered to students nearing graduation from an adult learning and development master’s degree program.
The article addresses the rate of reliability among the six factor groups and how the reliability measurements
change when specific items are deleted from the ALDSS Scale. The researchers measured the construct reliability
bias among the six factor groups with respect to race.
Keywords: Adult Education, Adult Learning and Development, Reliability, Graduate Students, Higher
Education, Student Satisfaction, Survey Development.
1. Introduction
We in the United States are currently living in an era whereby the common political view is to demand that
federal and state departments and agencies operate under strict fiscal principles. The current political trend has
left the field of higher education extremely vulnerable, with respect to funding of academic programs. Brand
(1997, 2000) predicted this demand for market driven higher education more than a decade ago when he saw the
signs that stemmed from the market driven policies being passed by the U.S. federal government. According to
Neal (1995), “American higher education faces a common but unwelcomed dilemma – to prove its worth to an
increasingly skeptical and critical public” (p. 5). When state legislatures discuss issues related to higher
education, Ruppert (1995) suggests that they frequently use “terms like accountability, performance, and
productivity” (p. 11).
The current political climate has also empowered many accreditation agencies in the United States, such as
NCATE, to oversee the academic accountability process for many colleges of education. Adult education
graduate programs are often assigned to university departments who may also include counseling, higher
education, school administration, and supervision (Merriam & Brockett, 2007). Therefore, it is rare to see a
stand-alone adult education graduate program in the United States. Since adult education graduate programs are
not licensure based programs, we are often left more vulnerable to scrutiny by the state departments of education
because we do not have an accrediting body. Adult education programs are often seen as less important or not
taken seriously as those programs that have NCATE or CACREP as their accreditation credential. However,
adult education graduate programs are not loosely run academic programs in higher education. In addition to
meeting the state and university mandates, adult education professors and academic programs are governed by a
well defined set of standards set forth by the Commission of Professors of Adult Education (CPAE, 2010).
Because CPAE is a non-accredited body, our standards are often viewed in the public and private sectors as adhoc at best – I might add in much of the same manner that the nine ethical responsibilities adult educators have
toward adult learners were formulated and described by Wood (1996).
1
The Special Issue on Social Science Research
© Centre for Promoting Ideas, USA
www.ijhssnet.com
The majority of the U.S. states in 1992 were experiencing budgetary problems due to the national economic
recession, which according to Ruppert (1995) led to many state colleges and universities receiving less funding to
run its academic programs for a period of two or more years. Between the demand for budgetary reconciliation
among many U.S. universities and the lack of academic identity among many adult education graduate programs,
the field has seen a number of adult education graduate programs, such as those at Boston University, Cornell
University, Ohio State University, Oklahoma State University, Syracuse University, and the University of Alaska
at Anchorage close since the late 1980s (Knox & Fleming, 2010).
Given the demand for more accountability on the part of U.S. colleges and universities as described by Sandmann
(2010), the faculty from an Adult Learning and Development (ALD) master’s degree program at a large state
university in the Midwest region of the U.S. decided to develop measures to evaluate their graduate program. The
ALD faculty knew that they could quickly generate the student enrollment and retention numbers for the
university administrators. The ALD faculty also felt comfortable in justifying to the university administrators the
level of adult learning which was occurring in the classroom. For example, all ALD students were required to
complete an exit strategy component, which consisted of students completing one of the following options: ALD
portfolio course, comprehensive exam, research project, or master’s thesis. Regardless of which exit strategy
option the students chose, the ALD faculty knew that the students would have to demonstrate the level of adult
learning that they had gained through their participation in the master’s degree program. The one area in which
the ALD faculty did not have a mechanism for program evaluation was in measuring student satisfaction.
Therefore, the researchers decided to develop the Adult Learning and Development Student Satisfaction Scale
(ALDSS Scale).
2. Purpose Statement
The one area in which the ALD faculty did not have a mechanism for program evaluation was in measuring
student satisfaction. Therefore, the researchers intend to briefly discuss the development of the Adult Learning
and Development Student Satisfaction Scale (ALDSS Scale) and the degree to which the factored groups warrant
reliable measurements. The purpose of this paper is to discuss the development of the ALDSS Scale and to
measure the reliability of the six factor groups with respect to the individual survey items. This investigation is
driven by the following three research questions:
1. Does each of the six factor groups warrant an acceptable rate of reliability?
2. Does the rate of reliability increase among any of the six factor groups when specific items are
removed?
3. Does each of the six factor groups warrant an acceptable rate of reliability with respect to race?
3. Literature Review
The researchers conducted an extensive literature review of previous studies measuring program satisfaction at
higher education institutions. Therefore, the researchers found only four empirical studies that address some form
of student satisfaction within colleges and universities (Lounsbury, Saudargas, Gibson, & Leong, 2005; Nauta,
2007; Spooren, Mortelmans, & Denekens, 2007; Zullig, Huebner, and Pun, 2009). In all four cases, the authors
primarily studied undergraduate students and only the Spooren, Mortelmans, and Denekens (2007) study made
note toward the psychometric values among the factor measures as it pertained to construct reliability. The
researchers did not find any student satisfaction studies that pertained to the evaluation of graduate programs.
The empirical study from Spooren, Mortelmans, and Denekens (2007) sought to identify the important factors that
influence the level of teacher quality in the higher education classroom. They developed a 31-item scale to
measure the level of teacher quality in the classroom. While using a factor analysis approaches, Spooren et al.
were able to identify 10 factor groupings in which to measure the level of teacher quality in the classroom. The
ten factor groups consist of (1) the clarity of objectives, (2) the value of subject matter, (3) the build-up of subject
matter, (4) the presentation skills, (5) the harmony organization course – learning process, (6) the contribution to
understanding the subject matter, (7) the course difficulty, (8) the help of the teacher during the learning process,
(9) the authenticity of the examination(s), and (10) the formative examination(s). All ten factor groups warranted
Chronbach alpha levels ranging between .73 and .88.
2
International Journal of Humanities and Social Science
Vol. 2 No. 14 [Special Issue - July 2012]
Figure 1. Macro-level Approach to Measuring Student Satisfaction in an Adult Learning and Development
Master’s Degree Program.
Six Factor Groups
for Program
Evaluation
While these ten factors were considered during the development of the ALDSS Scale, the overall focus of our
research was slightly different. In developing the ALDSS Scale, the researchers were interested in developing a
survey that would measure the students’ overall experience with the ALD master’s degree program. The ALDSS
Scale was designed to measure the level of student satisfaction with respect to six factor groups (see Figure 1).
The six factor groups include: (1) curriculum, (2) learning format, (3) course materials, (4) program access, (5)
faculty and instruction, and (6) faculty advising.
The curriculum, learning format, and course materials factor groups were designed to measure the students’
satisfaction with respect to theory and professional practice. The program access factor was designed to measure
the students’ satisfaction with respect to their ability to participate in the adult learning and development program
in relation to their work and family needs. The faculty and instruction factor was designed to measure the
satisfaction the students had with the delivery of information in the classroom with respect to their adult learning
needs. The faculty advising factor was developed to measure the students’ satisfaction with respect to their
academic advising sessions with the ALD faculty.
4. Research Methodology
The researchers will discuss the process by which this study was developed and administered in order to measure
the level of reliability between the six factor groups pertaining to student satisfaction. This section will discuss
three areas of research methodology, which include: the instrumentation, the data analysis, and the sample. The
researchers also will discuss these three areas of research methodology with respect to theory and practice.
4.1 Instrumentation
The researchers used brainstorming techniques to develop the initial survey. The initial item pool consisted of 60items. In crafting the survey items and rate scale, the researchers followed the instrument development methods
described by Converse and Presser (1986) and Spector (1992). The survey was pre-piloted and reviewed by four
faculty members in the area of adult education.
3
The Special Issue on Social Science Research
© Centre for Promoting Ideas, USA
www.ijhssnet.com
It was determined through the pre-pilot process that the survey consisted of items that could be categorized into
six distinct groups. In addition to sorting the items into distinct groups, the pre-pilot session allowed the
researchers to delete the redundant items. The researchers made a consistent attempt to develop a survey that
warranted an equal number of items for each distinct group. Therefore, the researchers were able to settle on a 6item scale for each of the six groups.
The ALDSS Scale consists of a 36-item scale, shown later in Table 1. The items for the ALDSS Scale are grouped
on the survey with respect to the following six factor groups: (1) curriculum, (2) learning format, (3) course
materials, (4) program access, (5) faculty and instruction, and (6) faculty advising. In addition, all of the items on
the ALDSS Scale are equally distributed among the six factor groups. For example, items 1-6 represent the
curriculum factor, items 7-12 represent the learning format factor, items 13-18 represent the course materials
factor, items 19-24 represent the program access factor, items 25-30 represent the faculty and instruction factor,
and items 31-36 represent the faculty advising factor.
The students were asked to rate each of the items on the ALDSS Scale using a 6-point Likert scale (Likert, 1932;
Likert, Roslow, & Murphy, 1934) ranging from 1 = Strongly Disagree and 6 = Strongly Agree. The researchers
chose to utilize a Likert scale for measuring student satisfaction because of its historical acceptance among
researchers who study attitudinal data among adult learners (e.g., Ferguson, 1941; Fraenkel & Wallen, 2003; Guy
& Norvell; 1977; Mueller, 1986; Spector, 1992; Spooren et al., 2007; Thurstone, 1928). One advantage in using a
Likert scale is that the items representing each of the factors can be converted into a scale mean statistic.
Therefore, each of the scale mean scores for each of the factor groups can be compared statistically with respect to
its measureable difference. According to Spooren et al. (2007), a second “advantage of Likert scales is the ability
to test each scale for reliability by means of the Cronbach’s alpha statistic” (p. 670). The 6-point rating scale was
chosen by the researchers, rather than the normal 5-point scale, because it requires the participants to rate either a
positive or negative response for each of the scale items.
4.2 Data Analysis
The item validation process occurred throughout the development of the ALDSS Scale. On two different
occasions, the researchers piloted the ALDSS Scale to five ALD students in order to test for the content validity
(Borg, Gall, & Gall, 1993) and construct validity (Carmines & Zeller, 1979) of the items with respect to the six
factor groups. Each of the 36-items was typed on a single strip of paper. The students were given all 36 strips of
paper and asked to read and place each item-strip into one of six envelopes representing the six factor groups.
According to Carmines and Zeller (1979, p. 27), “construct validation focuses on the extent to which a measure
performs in accordance with theoretical expectations.” Therefore, when items were placed into the wrong
envelope, the researchers consulted with the validation participants to better understand the way in which they
interpreted the item in question. The researchers either reworded the item or eliminated the item from the ALDSS
Scale. When an item was deleted for a particular factor group, the researchers developed a new item for the factor
group. The new item was based upon the recommendations provided by the ALD students participating in the
validation study. The validation process took two months to complete.
The data from the students who participated in the ALDSS Scale was entered into an SPSS 14.0 (SPSS Inc., 2005)
dataset. The rate of reliability among the six factor groups was measured using the internal consistency method.
The internal consistency method was employed using the Cronbach alpha coefficient (Cronbach, 1951).
According to Fraenkel and Wallen (2003) and Spector (1992), an acceptable Cronbach alpha coefficient for most
studies administered in educational settings is a reliability coefficient of .70 or higher. However, due to the
moderate sample size for this study, the researchers elected to use a Cronbach alpha coefficient of .80 or higher as
an acceptable rate of reliability.
While testing for the rate of reliability among the six factor groups, the researchers looked at the Cronbach alpha
coefficient for each factor when specific items are deleted from the study. When measuring the adjusted
Cronbach alpha coefficient for factors with deleted items, the SPSS software evaluates the inter-item correlation
within each of the six factor groups using a Pearson’s Correlation statistic to make the adjustment for the
Cronbach alpha coefficient. The researchers also used the Pearson’s Correlation statistic to measure the inter
correlation between the six factor groups.
4
International Journal of Humanities and Social Science
Vol. 2 No. 14 [Special Issue - July 2012]
Finally, the researchers used the Cronbach alpha coefficient to measure the reliability among the six factor groups
as it pertained to race. The researchers ran independent measurements for the White students and for the
Students-of-Color. The purpose for measuring the reliability coefficients with respect to race was to be able to
determine if the ALDSS Scale was free of any scale item bias (e.g., Huebner & Dew, 1993; Plant & Devine, 1998;
Tetlock & Mitchell, 2008). The researchers also would have desired to study the rate of reliability with respect to
gender, but the extreme differences in sample size between the male and female students prevented them from
looking at gender bias issues, as women were the majority population in the sample.
4.3 Sample
The sample size for this study consisted of 116 ALD graduate students nearing graduation, which warranted a
93.5% rate of response. It was believed that these students would provide a more reliable critique of the six
factors groups regarding their satisfaction with the ALD program than a student who was new to the academic
program. Whereas, a new student would not be able to rate some of the items on the ALDSS Scale. All student
participation was voluntary and their identities were held with strict confidentiality. The sample for this study
consists of a very diverse group of adult learners. The ALD students had a mean age of 38.4 years, who ranged
between 23-59 years. The sample also was a very experienced group of adult learners, whose level of
professional experience had a mean score of 11.0 years and a ranged between 0-36 years of experience. This
study consists of a gender skewed sample, with 89.7% of the participants (n=104) representing female adult
learners. In addition, the sample was racially diverse in that 53.4% of the adult learners represented students-ofcolor, with 50.0% African-American (n=58) and 3.4% Hispanic/Latino (n=4). The majority of the sample
(62.9%) was non-married adult learners (n=73), which consisted of those who were single (n=61), divorced
(n=11), and widowed (n=1). In addition, the majority of this sample (51.7%) did not have any dependent
children. Although, the adult learners who had dependent children ranged between 1-4 children, with the majority
consisting of adult learners with 1-child (n=23) and 2-children (n=26). Finally, fifty percent of the adult learners
(n=58) for this study had an annual salary that ranged between $30,000-$49,999, while 28.4% earned less than
$30,000 annually and 21.6% earned $50,000 or more annually. The sample also consisted of 11 adult learners
(9.5%) who had an annual salary below that of the national poverty level.
5. Research Findings
In this section of the paper, the researchers will discuss the results from testing the reliability measures for the six
factor groups and the strength of the individual items among the six factor groups. First, the researchers will
discuss the findings with respect to the first research question which measures the rate of reliability among the six
factor groups. The researchers will address the second research question that measures the degree to which each
item among the six factor groups either strengthens or weakens the reliability measurement for the respective
factors. Finally, the researchers will discuss the findings of the third research question which measures the rate of
reliability of the six factor groups with respect to the race of the sample.
5.1 Does each of the six factor groups warrant an acceptable rate of reliability?
The researchers measured the rate of reliability among the six factor groups using the Cronbach (1951) alpha
statistic. The reliability measurements for each of the six factor groups were generated from a sample size of 116
students (see Table 1). As suggested earlier, the researchers used the Cronbach alpha level of .80 or greater to
serve as the benchmark for determining the acceptable rates of reliability among the six factor groups.
The findings suggest that each of the six factor groups had Cronbach alpha levels greater than .80. The learning
format factor group warranted the highest rate of reliability, with an alpha level of .91 and a factor mean of 4.89.
The second highest rate of reliability measurement was found among the faculty advising factor group, with an
alpha value of .88 and a factor mean of 5.00. The course materials factor group had the third highest rate of
reliability statistic among the six factor groups, with an alpha value of .85 and a factor mean of 5.01. The
program access and curriculum factor groups warranted alpha values of .84 and .82 and factor means of 5.12 and
5.13 respectively. The factor with the lowest reliability rating, but still acceptable, was the faculty and instruction
factor group which warranted an alpha value of .81 and a factor mean of 5.31. One of the interesting statistics
that the researchers found was that while the factor means decreased, the rates of reliability increased between
each of the six factor groups.
5
The Special Issue on Social Science Research
© Centre for Promoting Ideas, USA
www.ijhssnet.com
5.2 Does the rate of reliability increase among any of the six factor groups when specific items are
removed?
The purpose of this question was to test whether item deletion on the ALDSS Scale resulted in an increase in the
rate of reliability among its respective factor group. As in the case of the first research question, this research
question was tested using a sample size of 116 students. The findings suggest that a decrease in the rate of
reliability would occur if any items were deleted among the curriculum (α = .78 to .81), learning format (α = .89
to .90), and program access (α = .80 to .82) factor groups. The findings suggest that the rate of reliability for the
faculty advising factor group would either remain the same or decrease, if any of the six items were deleted.
Deleting any of the six items would result in a Cronbach alpha range between .84 and .88. There were two items,
if deleted from the ALDSS Scale, would have resulted in the faculty advising factor group hold the current
Cronbach alpha value of .88. The first item in question was Item-31, which read “I was able to schedule
academic appointments with my advisor in a timely manner.” The second item in question was Item-34, which
read “My advisor suggested I take specific elective courses that would prepare me for professional practice.”
The findings suggest one item among the course materials factor group and one item among the faculty and
instruction factor group are called into question. With regards to the course materials factor group, if either one
of Items 13-17 are deleted, then this would result in a decrease in the rate of reliability (α = .80 to .82) from the
factor group’s current Cronbach alpha value of .85. However, if Item-18 (“The course syllabi were clear with
respect to the course objectives.”) was deleted from the course materials factor group, then this would increase
the rate of reliability to a Cronbach alpha value of .88. It is also important to note that the findings suggest that
Item-18 was the highest rated item among the course materials factor group, with an item mean score of 5.33. In
looking at the inter-item correlation data, the findings suggest that Item-18 warranted a Pearson’s correlation
statistic greater than a 95 percent level of confidence with respect to the five remaining items in the course
materials factor group.
With respect to the faculty and instruction factor group, if either one of Items 25-29 are deleted, then this would
result in a decrease in the rate of reliability (α = .74 to .78) from the factor group’s current Cronbach alpha value
of .81. In contrast, if Item-30 (“The ALD faculty would participate frequently in the web-based course
discussions for on-line course.”) was deleted from the faculty and instruction factor group, then this would
increase the rate of reliability to a Cronbach alpha value of .87. It is important to mention that the findings
suggest that Item-30 was the lowest rated item among the faculty and instruction factor group, with an item mean
score of 4.71. The findings from the inter-item correlation data suggest that Item-30 warranted a Pearson’s
correlation statistic greater than a 95 percent level of confidence with respect to the five remaining items in the
faculty and instruction factor group.
5.3 Does each of the six factor groups warrant an acceptable rate of reliability with respect race?
The purpose of this research question was to test whether there was any racial bias regarding reliability measures
among the six factor groups for the ALDSS Scale (see Table 2). This was accomplished by looking at the White
student sample and the Students-of-Color sample independently with respect to construct reliability testing.
According to the findings, the following four factor groups: (1) curriculum, (4) course materials, (3) program
access, and (4) faculty advising had greater Cronbach alpha scores, ranging between .84 and .90, among the
Students-of-Color sample than for the White student sample who had Cronbach alpha scores that ranged between
.79 and .87 for the same four factor groups. In the case of the learning format factor group, the findings suggest
that there was no difference in the reliability scores between the two sample groups, as both racial groups
warranted a Cronbach alpha score of .91, which was the same value for the total sample. In contrast to the other
factor groups, the findings suggest that the rate of reliability for the faculty and instruction factor group was much
greater among the White student sample (α = .85) than for the Students-of-Color sample (α = .68). In addition to
the faculty and instruction factor group not warranting an acceptable Chronbach alpha score of .70 among the
Students-of-Color sample, the findings also suggest that there is likely a strong racial bias among the items
measuring faculty and instruction in favor of the White student sample.
When the researchers looked further at the internal consistency of the six items measuring faculty and instruction,
the findings suggests again that if Item-30 (“The ALD faculty would participate frequently in the web-based
course discussions for on-line course.”) was deleted from the ALDSS Scale that the rate of reliability for this
factor group would increase for both sample groups.
6
International Journal of Humanities and Social Science
Vol. 2 No. 14 [Special Issue - July 2012]
The findings suggested that if Item-30 was deleted that the faculty and instruction factor group would warrant a
Cronbach alpha score of .77 among the Students-of-Color sample and a Cronbach alpha score of .90 among the
White student sample. Although, deleting Item-30 would increase the rate of reliability among both samples, the
difference in the Cronbach alpha scores for the faculty and instruction factor group only reduced from .17 to .13
points between the two samples. The findings also suggest that if any of the five remaining items (Items 25-29)
measuring faculty and instruction were deleted; this would result in a decrease in the Cronbach alpha score for
this factor group among both samples.
Table 1. The Measurement of Reliability for the Six Factor Groups and the Influence of Item Deletion
(N=116).
I.
Curriculum:
Factor Mean = 5.13
SD = 0.81
Cronbach Alpha = .82
Current
Items
Item Mean
1-The core ALD courses (ALD 605, 607, 663, 645, and 688) increased my knowledge of
the adult learning and development theories.
5.47
2-The core ALD courses were applicable to my professional practice.
5.07
3-The elective courses were applicable to my professional practice.
4.99
4-The elective courses represented my areas of interests in professional practice.
4.96
5-The ALD internship course increased my skills for professional practice.
5.23
6-The ALD internship course increased my theoretical framework with respect to my
professional practice.
5.04
II. Learning Format: Factor Mean = 4.89 SD = 0.86 Cronbach Alpha = .91
Current
Items
Item Mean
7-The course lectures were applicable to my professional practice.
4.84
8-The in-class group discussions were applicable to my professional practice.
5.02
9-The group projects were applicable to my professional practice.
4.84
10-The course papers were applicable to my professional practice.
5.01
11-The student presentations were applicable to my professional practice.
4.78
12-The course media (e.g., films, videos, and other productions) were applicable to my
professional practice.
4.85
III. Course Materials: Factor Mean = 5.01 SD = 0.72 Cronbach Alpha = .85
Current
Items
Item Mean
13-The textbooks increased my knowledge of the adult learning and development
theories.
5.04
14-The textbooks were applicable to my professional practice.
4.79
15-The additional course readings increased my knowledge of the adult learning and
development theories.
5.14
16-The additional course readings were applicable to my professional practice.
4.84
17-The course handouts were applicable to my professional practice.
4.91
18-The course syllabi were clear with respect to the course objectives.
5.33
IV. Program Access:
Factor Mean = 5.12
SD = 0.92
Cronbach Alpha if
Item Deleted
.81
.78
.78
.81
.78
.79
Cronbach Alpha if
Item Deleted
.89
.90
.89
.89
.90
.90
Cronbach Alpha if
Item Deleted
.82
.81
.81
.80
.82
.88*
Cronbach Alpha = .84
Items
19-The ALD courses were offered at a time of day that did not conflict with my work
schedule.
20-The ALD courses were offered at a time of day that allowed me to work around my
family needs.
21-The ALD core courses were offered frequently enough to meet my academic
requirements.
22-The elective ALD courses were offered frequently enough to meet my academic
needs.
23-The web-based ALD courses allowed me to participate in coursework that did not
conflict with my work schedule.
24-The web-based ALD courses allowed me to participate in coursework that did not
conflict with my family needs.
Current
Item Mean
Cronbach Alpha if
Item Deleted
5.16
.81
5.19
.80
4.96
.81
4.51
.82
5.51
.81
5.40
.81
*Denotes an increase in the Cronbach alpha value for a factor if a specific item is deleted.
7
The Special Issue on Social Science Research
© Centre for Promoting Ideas, USA
www.ijhssnet.com
Table 1. The Measurement of Reliability for the Six Factor Groups and the Influence of
Item Deletion (N=116). (Continued)
V.
Faculty and Instruction:
Factor Mean = 5.31
SD = 0.71
Cronbach Alpha = .81
Current
Items
Item Mean
25-The ALD faculty had a strong understanding of the course topics.
5.62
26-The ALD faculty stimulated my interests in the field of adult learning and
development.
5.44
27-The ALD faculty clearly stated the requirements for each course.
5.52
28-The ALD faculty offered multiple modes of learning that met my adult learning
needs.
5.32
29-The ALD faculty provided a quality critique of my coursework.
5.25
30-The ALD faculty would participate frequently in the web-based course discussions
for on-line courses.
4.71
VI. Faculty Advising: Factor Mean = 5.00 SD = 1.06 Cronbach Alpha = .88
Current
Items
Item Mean
31-I was able to schedule academic appointments with my advisor in a timely manner.
5.18
32-My advisor responded to telephone/email messages in a timely manner.
5.16
33-My advisor effectively explained the requirements for the ALD master’s degree
program.
5.18
34-My advisor suggested I take specific elective courses that would prepare me for
professional practice.
4.55
35-My advisor effectively explained the requirements for the ALD internship course.
4.78
36-My advisor effectively explained the ALD exit strategy requirement.
5.16
Cronbach Alpha if
Item Deleted
.77
.75
.78
.76
.74
.87*
Cronbach Alpha if
Item Deleted
.88
.87
.84
.88
.86
.85
*Denotes an increase in the Cronbach alpha value for a factor if a specific item is deleted.
Table 2. The Measurement of Reliability for the Six Factor Groups with Respect to Race (N=116).
Factor Groups
I.
Curriculum
II.
Learning Format
III.
Course Materials
IV. Program Access
V.
Faculty and Instruction
VI. Faculty Advising
Total Sample
Cronbach Alpha
(N=116)
.81
.91
.85
.84
.81
.88
White Students
Cronbach Alpha
(n=54)
.79
.91
.84
.80
.85
.87
Students-of-Color
Cronbach Alpha
(n=62)
.84
.91
.86
.86
.68
.90
6. Discussion
In reflecting upon the findings of this study, the researchers believe that there are a number of positive outcomes
due to the development of the ALDSS Scale. One of the most obvious aspects to occur is that this study suggests
that all six factor groups (curriculum, learning format, course materials, program access, faculty and instruction,
& faculty advising) for the ALDSS Scale were found to be reliable factors. Secondly, the most important outcome
stemming from the development of the ALDSS Scale is that the adult learning and development faculty have a
reliable mechanism for measuring how satisfied their students are with the ALD master’s degree program. The
ALDSS Scale has become a very powerful tool in that the ALD faculty now has a working model (see Figure 1)
for measuring ALD program’s level of success with respect to numerous macro-level factors. Often when we
think of program evaluation in higher education, we tend to look independently at the numerous micro-level
factors, such as: teaching methods, adult learning gains, grade-point averages, graduation rates, graduate
employment rates, and total student credit hours generated (e.g., enrollment rates and tuition revenue). While
these are valid factors to study, most of the factors are not directly influenced by the university faculty. In
contrast, the researchers developed the ALDSS Scale to measure six factors that summarized the overall
administration of the ALD master’s degree program and in which the faculty had a great deal of control over the
total administration of these factors.
8
International Journal of Humanities and Social Science
Vol. 2 No. 14 [Special Issue - July 2012]
In essence, the ALDSS Scale is as much of a measurement tool for reflecting faculty performance, as it is for
program evaluation. While the ALDSS Scale is a reliable instrument, this is not to say that the ALDSS Scale is
without its limitations. For example, the findings suggested that two of the six factor groups had one item (Item18 and Item-30) that reduced the potential rate of reliability for the factor groups by .03 and .06 points
respectively. The course materials factor and the faculty and instruction factor did have reliability rates above
.80, but anytime specific items are shown to reduce the rate of reliability by .03 points or more, one must take a
second look at these items.
With regard to Item-18 for the course materials factor (“The course syllabi were clear with respect to the course
objectives.”), the researchers believe that this is a valid item with respect to its application to the factor group.
The researchers’ belief was reassured from the findings from the Pearson’s correlation, which suggested that there
was a statistically significant correlation between Item-18 and the other items that represent the course materials
factor group. In addition, Item-18 was the highest rated item among the course materials factor. The problem
regarding Item-18 is not with its validity, but with its language. The researchers believe that the words “clear”
and “objectives” leave the students with a vague understanding of Item-18. Therefore, the researchers
recommend changing Item-18 to read, “The course syllabi identified the course requirements.”
With regard to Item-30 for the faculty and instruction factor (“The ALD faculty would participate frequently in the
web-based course discussions for on-line courses.”), the researchers also believe that this is a valid item with
respect to its application to the factor group. As in the case for Item-18, the findings from the Pearson’s
correlation suggested that there was a statistically significant inter-item correlation between Item-30 and the other
items that represent the faculty and instruction factor group. Therefore, to immediately delete this item would
pose a problem of appearing to skew the results, because Item-30 was by far the lowest rated item among the
faculty and instruction factor group. The researchers on the surface believed that the phrase “would participate
frequently” causes Item-30 to read less clear. Therefore, the researchers are initially recommending that Item-30
be reworded to read, “The ALD faculty participated in the course discussions for on-line courses.” However,
when the researchers sought to test the third research question that looked at whether there was any racial bias
among the items for the six factor groups, a more serious problem developed with respect to Item-30.
When the researchers retested the Cronbach alpha levels among the six factor groups independently with respect
to race, the findings suggested that five of the factor groups warranted greater Cronbach alpha levels among the
Students-of-Color sample than for the White student sample. This was a very important finding, because it
illustrated that there was no bias with respect to race among the items that measure the following five factor
groups: curriculum, learning format, course materials, program access, and faculty advising. Given that the
majority of the ALD graduate students are Students-of-Color, this finding suggest that when we were developing
the items for the ALDSS Scale that we were at least cognizant of the varying cultural needs and interests with
respect to our adult learners.
While these positive findings regarding construct racial bias were exciting news, we cannot ignore the negative
findings that suggest that there was a strong racial bias with respect to the faculty and instruction factor group
items in favor of the White student sample (e.g., Huebner & Dew, 1993; Plant & Devine, 1998; Tetlock &
Mitchell, 2008). The findings suggested that if Item-30 were deleted, then the faculty and instruction factor group
would warrant an acceptable Cronbach alpha level among the Students-of-Color sample as well as further
increase the already acceptable Cronbach alpha level for the White student sample. Therefore, given the racial
bias issue with regards to this item, the researchers recommend that Item-30 be definitely dropped from the
ALDSS Scale, regardless of any previous concerns that the researchers may have had regarding the deletion of this
item. Nevertheless, a silent issue still hovers over the faculty and instruction factor group. While we can solve
the unacceptable Cronbach alpha level for the faculty and instruction factor group among the Students-of-Color
sample by deleting Item-30, we have only decreased the disparity between the two sample groups’ Chronbach
alpha levels by a total of .03 points. Therefore, there is still a .13 Cronbach alpha point difference between the
reliability scores between the two racial groups, in favor of the White student sample.
When looking at the five remaining items (Items 25-29) measuring faculty and instruction issues, it is not visually
apparent to the researchers as to what is influencing the racial bias with respect to the language of the five items.
However, the researchers acknowledge that they are evaluating the items from our White lenses, as both
researchers are not Faculty-of-Color.
9
The Special Issue on Social Science Research
© Centre for Promoting Ideas, USA
www.ijhssnet.com
Because the faculty and instruction factor group shows an acceptable rate of reliability (following the deletion of
Item-30) for both sample groups, the researchers recommend that the five items (Items 25-29) measuring the
faculty and instruction factor group remain within the ALDSS Scale. However, the researchers plan to conduct
further statistical measurements to depict additional psychometric flaws with respect to the racial bias of these
five items. Furthermore, the greatest identification of racial bias among the five items may only occur from an
extensive qualitative investigation with ALD Students-of-Color. It is through their lenses that the missing pieces
of the puzzle can be found.
According to Sandmann (2010), it is becoming more apparent that the faculty members who teach in adult
education programs at the college or university level identify ways in which they can justify their academic
programs’ accomplishments and worth to the community. Therefore, the researchers believe that faculty and
administrators in adult education graduate programs can easily transfer many of the same factors and items from
the ALDSS Scale to their own academic program with respect to measuring student satisfaction. In fact, the intercorrelation data suggest that the six factor groups, which make up the ALDSS Scale, are statistically significant
factors for measuring student satisfaction.
Furthermore, this investigation illustrates the need for survey developers to be cognizant of specific biases when
developing psychometric instruments. Too often when we develop questionnaires for quantitative research, we
tend to look only at the surface with respect to the reliability of the measured constructs. Had the researchers not
delved further into the data, they would not have discovered the racial bias among one of the factor groups on the
ALDSS Scale. The researchers only regret not having a more inclusive sample so that they could have studied the
potential gender bias among the six factor groups.
The researchers’ goal is that the ALDSS Scale will provide adult and higher education practitioners with another
means for evaluating the success of their academic programs. With the current era of higher education bound by
program accountability, adult education graduate programs must be prepared to justify their existence to the
decision makers within higher education. In successfully providing valid program evaluation ideas, perhaps the
closing of adult education graduate programs by universities around the country can be stymied or averted. As
Sandmann (2010, p. 228) suggests, the call “for accountability and transparency will continue to shape higher
education.” Milton, Watkins, Studdard and Burch (2003) urges professional adult educators to pay attention to
the larger context in which their programs exist, including the political climate, budget and resource constraints,
and competing programs. As adult educators and program planners, it is important to be attuned the changes that
are shaping higher education and thus graduate adult education programs.
Finally, the researchers encourage all faculty members to understand the importance of program evaluation and
conduct further empirical research to further program evaluation efforts in adult education graduate programs.
This is one area of study which is underserved yet vital to the continuing successful existence of adult education
graduate programs within the higher education arena. Future research and discussion should further examine how
to strengthen adult education graduate programs in the turbulent political context in which they exist.
7. References
Borg, W. R., Gall, J. P., & Gall, M. D. (1993). Applying educational research: A practical guide (3rd ed.). White
Plains, NY: Longman.
Brand, M. (1997). Some major trends in higher education. An address delivered by the President of Indiana
University to the Economic Club of Indianapolis, Indiana on February 7.
Brand, M. (2000). Changing faculty roles in research institutions using pathways strategy. Change, 32(6), 42-45.
Carmines, E. G., & Zeller, R. A. (1979). Reliability and validity assessment. Sage University
Paper series on the Quantitative Applications in the Social Sciences, series no. 17. Thousand Oaks, CA: Sage.
Converse, J. M., & Presser, S. (1986). Survey questions: Handcrafting the standardized questionnaire. Sage
University Paper series on the Quantitative Applications in the Social Sciences, series no. 63. Thousand
Oaks, CA: Sage.
CPAE (2010). Standards for graduate programs in adult education. Bowie, MD: Commission of Professors of
Adult Education: American Association for Adult and Continuing Education.
http://web.memberclicks.com/mc/page.do?sitePageId=85438&orgId=cpae
Cronbach, L. J. (1951). Coefficient alpha and the internal structure of tests. Psychometrika, 16, 297-334.
10
International Journal of Humanities and Social Science
Vol. 2 No. 14 [Special Issue - July 2012]
Ferguson, L. W. (1941). A study of the Likert technique of attitude scale construction. Journal of Social
Psychology, 13, 51-57.
Fraenkel, J. R., & Wallen, N. E. (2003). How to design and evaluate research in education (5th ed.). New York:
McGraw-Hill.
Guy, R. F., & Norvell, M. (1977). The neutral point on a Likert scale. Journal of Psychology, 95, 199-204.
Huebner, E. S., & Dew, T. (1993). An evaluation of racial bias in a life satisfaction scale. Psychology in the
Schools, 30, 305-309.
Knox, A. B. & Fleming, J. E. (2010). Professionalization of the field of adult and continuing education (pp. 125134). In C.E. Kasworm, A.D. Rose, & J. M. Ross-Gordon (Eds.) Handbook of Adult & Continuing
Education (2010 Edition). Thousand Oaks: SAGE.
Likert, R. (1932). A technique for measurement of attitudes. Archives of Psychology, 22(140), 55.
Likert, R., Roslow, S., & Murphy, G. (1934). A simple and reliable method of scoring the Thurstone attitude
scales. Journal of Social Psychology, 5, 228-238.
Lounsbury, J. W., Saudargas, R. A., Gibson, L. W., & Leong, F. T. (2005). An investigation of broad and narrow
personality traits in relation to general and domain-specific life satisfaction of college students. Research
in Higher Education, 46, 707-729.
Merriam, S. & Brockett, R. (2007). The profession and practice of adult education. An introduction. San
Francisco: Jossey-Bass.
Mueller, D. J. (1986). Measuring social attitudes: A handbook for researchers and practitioners. New York:
Teachers College, Columbia University.
Milton, J., Watkins, K. E., Studdard, S. S., & Burch, M. (2003). The ever widening gyre: Factors affecting
change in adult education graduate programs in the United States. Adult Education Quarterly 54 (1), 2341.
Nauta, M. M. (2007). Assessing college students’ satisfaction with their academic majors. Journal of Career
Assessment, 15, 446-462.
Neal, J. E. (1995). Overview of policy and practice: Differences and similarities in developing higher education
accountability. In G. H. Gaither (Ed.), Assessing Performance in an Age of Accountability: Case Studies.
New Directions for Higher Education (pp. 5-10), No. 91. San Francisco, CA: Jossey-Bass.
Plant, E. A., & Devine, P. G. (1998). Internal and external motivation to respond without prejudice. Journal of
Personality and Social Psychology, 75, 811-832.
Ruppert, S. S. (1995). Roots and realities of state-level performance indicator systems. In G. H. Gaither (Ed.),
Assessing Performance in an Age of Accountability: Case Studies. New Directions for Higher Education
(pp. 11-23), No. 91. San Francisco, CA: Jossey-Bass.
Sandmann, L. R. (2010). Adults in four-year colleges and universities: Moving from the margin to mainstream?
In C. E. Kasworm, A. D. Rose, & J. M. Ross-Gordon (Eds.), Handbook of Adult and Continuing
Education (2010 ed.) (pp. 221-230). Thousand Oaks, CA: Sage.
SPSS Inc. (2005). SPSS 14.0 software. Chicago, IL: Author.
Spector, P. E. (1992). Summated rating scale construction: An introduction. Sage University Paper series on the
Quantitative Applications in the Social Sciences, series no. 82. Thousand Oaks, CA: Sage Publications.
Spooren, P., Mortelmans, D., & Denekens, J. (2007). Student evaluation of teaching quality in higher education:
Development of an instrument based on 10 Likert-scales. Assessment & Evaluation in Higher Education,
32, 667-679.
Tetlock, P. E., & Mitchell, G. (2008). Calibrating prejudice in milliseconds. Social Psychology Quarterly, 71, 1216.
Thurstone, L. L. (1928). Attitudes can be measured. American Journal of Sociology, 33, 529-554.
Wood, G. S., Jr. (1996). A code of ethics for all adult educators. Adult Learning, 8(2), 13-14.
Zullig, K. J., Heubner, E. S., & Pun, S. M. (2009). Demographic correlates of domain-based life satisfaction
reports of college students. Journal of Happiness Studies, 10, 229-238.
11
Download