Managing “At risk” business students: Statistical analysis of student data profiles Abstract

advertisement
Managing “At risk” business students:
Statistical analysis of student data profiles
Christo Potgieter
Wilfred Greyling
Vida Botes
Wintec
Wintec
Wintec
christo.potgieter@wintec.ac.nz
wilfred.greyling@wintec.ac.nz
vida.botes@wintec.ac.nz
Abstract
In this paper a multivariate model is developed to predict success or failure, helping to identify
potentially at-risk students, based on the first two assessments. A robust discriminant or regression
function will allow identification of students who obtained scores on these assessments that would put
them at risk of failing the course. This paper uses stepwise discriminant analysis and multiple
regression analysis to explore the data profiles for two closely related degree-level qualifications in
Business Studies at the same institution. Evidence for the Business Studies qualifications suggests
that a robust prediction model is possible. Therefore one could reflect on implications for more
efficient teaching and assessment policies and practices. In line with the theme of the NZABE
conference to question fundamentals”, one could ask whether exams are really required?! Although
various kinds of changes seem possible, there are many pitfalls that still need to be explored.
Key words: Student outcomes, completions, pass rates, examination, predictors, predictive efficiency
Literature review
The Tertiary Education Commission (TEC) has clearly stated that tertiary institutions have to improve
student outcomes (Tertiary Education Strategy 2007, Tertiary Education Strategy Monitoring
Information 2008, TEC to publish educational performance information 2010). Recently the NZ
Herald (2010) announced that government funding for institutions would be adjusted on the basis of
student outcomes. Tertiary institutions, it has been reported, need to set targets in their business plans
and report on achievement, specifically student pass rates and progression.
At WINTEC, the executive refers to a five-year window period to pursue and build a
reputation for “Quality and Student-Centeredness” (Our Direction, 2009). One of the eight key
strategic themes is “Quality and Outcomes”, stating amongst other things, that the institution will
pursue “high levels of student satisfaction and completion rates”. The institute’s commitment is
reflected in its Academic Direction (2010) – “Student retention, especially in the early stage of their
study programme, is a key strategy in improving student completions. We will improve this by having
a more explicit and integrated school and support service approach to retention and ensuring we have
the processes in place to monitor and report on progress made.”
Apart from attendance monitoring as a specific strategy among many to manage “at risk”
students, vocational tutors and Programme Managers generally act on results from assessments. It is
normal practice for tutors to pay more attention to students who fail assessments. It is generally
believed that the final outcome of student performance (pass/fail and final mark at end of semester)
can still be changed after the early assessments. Special intervention per student after the first
assessment then seems indicated and beneficial. In order to get the benefits of early intervention,
learners’ academic progress in the first half of the course ought to be scrutinised very closely.
Of course, if reliable information is available about student performance over the first half of
a programme, sustained intervention could be planned, avoiding “last minute crisis” attempts to
support students to pass. The assumption is that it is still possible for the student to pass in spite of
having failed a significant portion of the assessments in the semester! One could be tempted to argue
on the basis of personal experience that attempting to change the outcome for any of these at-risk
students is a lost cause – “If you fail at the start, you fail at the end”. If evidence of such fatalistic
outcomes exists, a case might be formulated against interventions because they might be poor return
on investment and a waste of resources.
A significant danger is that tutors may adopt the view, at a certain stage of the semester
(“generally when students have crossed the point of no return”), that students are “doomed to fail”
because, as tutors, they are no longer able to influence the outcome. However, the notion of a “selffulfilling prophecy” would only be relevant if there was actually still a chance of success, i.e. if
students who fail initially, succeed in the end. While there would always be exceptions to the rule, the
question is how many students fit this category. With this information the return on investment for
interventions might be explored, which is important, considering financial pressures experienced by
tertiary institutions.
So, it would be useful to have some insight into patterns of student progress, and how initial
assessments allow us to predict the likely outcomes. However, focusing on historical data only would
be limiting and reductionist, reflection on current intervention practices makes sense. This is done in a
section close to the end of the paper. At this stage it seems prudent to explore existing data. Since
most tutors implement some form of student intervention, assessment results and final outcomes
already reflect the effect of historical (likely current) practices. These practices need to be captured
too.
Research questions
Before formulating research questions, the assessment structures of all the modules in the two
qualifications were studied. Modules with assessment structures that deviated from the pattern of two
initial formative assessments and a final examination mark were then excluded. Thus, the criterion
for inclusion was that a module should have at least two initial assessments and a final examination
scores. With the targeted modules identified, the following research questions were formulated:
1. To what extent do the first two assessments in a module allow us to predict final course outcome?
2. How accurately can one predict the final outcome after each of the first and second assessments?
3. How significant in size is the number of students whose final results differ from initial assessment
scores?
4. Is the level of incorrect predictions from a specific statistical model acceptably low?
5. To what extent do the initial two assessments discriminate between the selected two
qualifications?
The project also offered the opportunity to consider the potential of the statistical package and its
tools to assist us in designing prediction models for future use by management. The software package
SPSS provides a powerful framework, is used widely and is available locally; for these reasons, the
package became the vehicle for analysing student data. This research project was an opportunity to
identify patterns in the student data that would promote wider institutional self-reflection. For this
reason, consistency with the premises of exploratory case studies (Yin, 1993) were considered,
prompting exploring programme-related and systemic issues.
Independent skilled observation is important for objective research so possible personal interest
ought to be recorded. The authors of this paper have personal interest in the data as managers and
have some training but no expertise in the advanced statistical methods used. For this reason data
analysis and interpretation are written (like case studies often are) in the first person as personal
dialogue showing reasoning of the participants.
Data processing
Our aim was to use existing organisational records and data sets to develop a statistical model for
predicting learner success on the basis of three independent variables (i.e. Assessments 1 and 2, as
well as Qualification). We argued that these variables could inform decisions to support learners at
different stages of their courses in the qualifications included in the study. Put differently, we
reasoned that some or all of these variables could be used as predictors of learner success. Once the
predictive efficiency of the variables in the model was established, we would be in a position to target
vulnerable learners for support to retain them in courses.
In summary, our statistical analysis explored to what extent the three predictors (Assessments
1 and 2, as well as Qualification) explained the variance on the classification variable, Pass/Fail, in a
discriminant analysis, or the variance on the dependent variable Final Mark in a multiple regression
analysis. Our purpose was to develop a statistical model that would allow us to provide at-risk
learners with appropriate and timeous support.
Student performance on the School of Business’s degree level programmes BBS and GradDip
was analysed.
These students are distributed over a wide range of modules within the two
qualifications. We labelled both the qualifications (BBS = 1; GradDip = 2) and each module in the
database so that in the longer term we could focus on student performance on configurations of
modules within these qualifications. All of the GradDip students already have degrees. Most
students on GradDip are from overseas (mostly India) while a significant number of international
students (average about 40%) are registered for BBS – for purposes of this paper, we do not
distinguish between domestic and international students. We dealt with the full cohorts registered for
the two qualifications to establish whether specific patterns would emerge for the full cohort from the
analysis.
We extracted student performance data from the institute’s student database for 2009,
covering semesters one and two, as well as Winter/Summer Schools. The latter were implemented in
2009 mostly for students from China under pathway agreements. The data were processed, which
meant that we retained only those student data sets where scores for Assessment 1, Assessment 2 and
Final Result (Pass or Fail) had been recorded.
Analysis
To respond to the research questions, two statistical procedures were used: a stepwise
discriminant analysis and a stepwise linear regression.
The stepwise discriminant analysis was
deemed appropriate because the predictors were used to predict membership of two discrete
categories (Pass=1 Fail=0). In this procedure, the variables are entered in sequence. When a specific
predictor is entered, its contribution to the model is assessed. If the predictor explains a significant
proportion of the variance on the classification variable, it is retained in the model. When a variable is
retained in the model, the variables that are already included in the model are again assessed to see
whether they continue to explain a significant proportion of the variance. If a variable in this
emerging configuration of variables no longer explains a significant proportion of the variance, it is
discarded. The end result is that this iterative process will yield the smallest, yet most efficient set of
predictor variables.
The same iterative process applies to the stepwise linear regression. However, the difference
is that this particular procedure yields a set of variables to predict scores on a continuous variable (in
this case, Final Mark).
Findings
First, we report the relevant descriptive statistics for the variables.
In Table 1 are the
descriptive statistics for the continuous variables measured in the assessments and the final mark:
Table 1: Descriptive Statistics for group
Mean
Std. Deviation N
Fin_M
64.827
10.4930
1790
Ass1_M 61.479
17.4199
1790
Ass2_M 68.493
14.5925
1790
Interestingly, the means for the scores on the two assessments and final mark vary from 61.5
% to 68.5 %. The means are clustered together within a narrow 6% range which suggests that both
the assessment practices in these programmes and student outcomes have been consistent across
assessment opportunities.
In Table 2, we report the classification table produced by the stepwise discriminant analysis.
It is clear from the table that the three variables allow us to classify the 83.3 % of the cohort of
students correctly within either the pass or the fail categories. Thus, 299 out of 1790 students were
classified incorrectly when the current prediction model, based on Assessments 1 and 2, as well as
Qualification, was used:
Table 2: Classification Results
Predicted
PF category
Fail
Pass
Actual
Fail
74
10
Pass
289
1417
%
Fail
88.1
11.9
Pass
16.9
83.1
(a. 83.3% of original grouped cases correctly classified)
Total
84
1706
100.0
100.0
Ideally one would want a more efficient model to predict membership of pass/fail so that false
positives and false negatives may be reduced. The false negatives (n = 289) are predicted negative,
but turn out positive. Perhaps the implication is that existing support measures and delivery strategies
aimed at promoting throughput are indeed effective. The false positives (n = 10) are predicted
positive, yet turn out negative. In this study these were a negligible proportion of the whole.
Prediction model
The findings suggest a relatively efficient start in selecting predictors for a prediction model.
The findings indicate that model 2, using the two assessments as predictors, explains 52.5 %
of the variance on the dependent variable Fin_Mark. The findings are reported in Table 3:
Table 3: Model Summaryc
Adjusted
Model R
R Square Square
1
.609a
.371
.371
2
.725b
.525
.525
a. Predictors: (Constant), Ass1_M
b. Predictors: (Constant), Ass1_M, Ass2_M
c. Dependent Variable: Fin_M
R Std. Error of
the Estimate
8.3244
7.2351
Change Statistics
R Square
Model
Change F Change
1
.371 1054.496
2
.154
579.898
c. Dependent Variable: Fin_M
df1
1
1
df2 Sig. F Change Durbin-Watson
1788
.000
1787
.000
1.249
In addition, the ANOVA results in Table 4 show that the F value signals a significant difference
between the P/F groups if they are compared on the basis of the three variables included in the
prediction model:
Table 4: ANOVAc
Model
Sum of Squares
Df
Mean Square
1
Regression
73072.060
1
73072.060
Residual
123900.744
1788
69.296
Total
196972.804
1789
2
Regression
103428.166
2
51714.083
Residual
93544.638
1787
52.347
Total
196972.804
1789
a. Predictors: (Constant), Ass1_M
b. Predictors: (Constant), Ass1_M, Ass2_M
c. Dependent Variable: Fin_M
F
1054.496
Sig.
.000a
987.903
.000b
Beta and T values
In table 5, the standardized Beta coefficients indicate how each variable contributes to the prediction
model. The Beta coefficient signals the number of standard deviations a dependent variable (in our
case, Fin_mark), will change when the value of a predictor variable changes by a standard deviation.
Thus, Ass2M and Ass1M have a significantly greater impact than Qualification. In fact, variable
Qual has been excluded from the model because its contribution to explaining the variance on Final
Mark is negligible. For the t values, the higher the value the more significant the impact of the
variable in predicting learner scores on the dependent variable, Fin_Mark.
Table 5: Coefficientsa
Model
Standardized
Unstandardized Coefficients Coefficients
B
Std. Error
Beta
1
(Constant)
42.272
.722
Ass1_M
.367
.011
.609
2
(Constant)
27.325
.883
Ass1_M
.270
.011
.449
Ass2_M
.305
.013
.424
a. Dependent Variable: Fin_M
T
58.555
32.473
30.959
25.482
24.081
Coefficientsa
Model
95.0% Confidence Interval for
B
Correlations
Lower Bound Upper Bound Zero-order Partial
1
(Constant)
40.856
43.688
Ass1_M
.345
.389
.609
.609
2
(Constant)
25.594
29.056
Ass1_M
.249
.291
.609
.516
Ass2_M
.280
.330
.594
.495
a. Dependent Variable: Fin_M
Coefficientsa
Model
Collinearity Statistics
Tolerance
VIF
1
(Constant)
Ass1_M
1.000
1.000
2
(Constant)
Ass1_M
.857
1.167
Ass2_M
.857
1.167
a. Dependent Variable: Fin_M
Excluded Variablesc
Model
Partial
Beta In
t
Sig.
Correlation
1
Ass2_M
.424a
24.081
.000
.495
Qual
-.002a
-.085
.932
-.002
2
Qual
.011b
.660
.509
.016
a. Predictors in the Model: (Constant), Ass1_M
b. Predictors in the Model: (Constant), Ass1_M, Ass2_M
Sig.
.000
.000
.000
.000
.000
Part
.609
.415
.393
Table 5: Coefficientsa
Model
Standardized
Unstandardized Coefficients Coefficients
B
Std. Error
Beta
1
(Constant)
42.272
.722
Ass1_M
.367
.011
.609
2
(Constant)
27.325
.883
Ass1_M
.270
.011
.449
Ass2_M
.305
.013
.424
c. Dependent Variable: Fin_M
T
58.555
32.473
30.959
25.482
24.081
Sig.
.000
.000
.000
.000
.000
Correlations
Table 6 presents correlations between the dependent and independent variables. These results
show moderate correlations between the two predictor variables and final mark. The correlation
between qualification and final mark is a negligible 0.025.
Table 6: Correlations
Pearson Correlation Fin_M
Ass1_M
Ass2_M
Qual
Sig. (1-tailed)
Fin_M
Ass1_M
Ass2_M
Qual
N
Fin_M
Ass1_M
Ass2_M
Qual
Fin_M Ass1_M Ass2_M
1.000
.609
.594
.609
1.000
.378
.594
.378
1.000
.025
.044
-.012
.
.000
.000
.000
.
.000
.000
.000
.
.144
.032
.299
1790
1790
1790
1790
1790
1790
1790
1790
1790
1790
1790
1790
Qual
.025
.044
-.012
1.000
.144
.032
.299
.
1790
1790
1790
1790
Results
Using stepwise regression analysis and stepwise discriminant analysis, the basic elements of a
model for predicting learner performance have been identified. This finding is supported by the
significance of the F value reported above (F =987.903, p = 0.000, Adjusted R2 = 0.525). The
predictors found to be significant in the model were (see percentages in tables): Assessment 2 scores
(Beta=0.449, p=0.000) and Assessment 1 scores (Beta=0.422, p=0.000), with Qualification excluded
from the model due to its negligible predictive efficiency. The Beta value explains the extent that the
classification variable will change when the predictor value changes by a standard deviation. The
large t-values in the table show that changes in these variables impact significantly on Final Mark.
Considerations
We felt that several interesting observations could be made about the statistical findings
reported above. For example, although GradDip and BBS share several classes, the patterns of
student performance do not reveal significant differences when these qualifications are compared.
This is interesting considering the GradDip students have already completed several years of
undergraduate study.
This is also valuable to know since each qualification has a significant
proportion of overseas students from two international regions in particular, namely India and China.
Although the means of the independent variable (Final Mark [65%]) and the assessment
variable means [69% and 62%] occur within a narrow band, we have little evidence to suggest that a
dramatic improvement in student performance did (or did not) occur. We could begin to define
specific metrics-based outcomes. For example, we could argue that a significant improvement would
be if the group scores improved by the standard deviation for the first assessment. We could argue
that the mean of 61.5% is the starting point, and if we add the standard deviation of 17.5, we have a
target: a mean of 79%. Or if we take half the standard deviation of 8.8, we could perhaps set a more
achievable objective of 70.5% [61.5 + 8.8 = 70.3 rounded off to 70.5]. We could then direct our
strategies at achieving SMART objectives (Specific, Measurable, Attainable, Relevant, Time-bound).
We have to note though that students who withdrew halfway after possibly failing early on, were not
included in the analysis. Thus, their performance data are not reflected in the coded data set we had
extracted for analysis.
A selection of modules which did not follow the same pattern of assessments was excluded
from the analysis. These included modules where two assessments were not administered; these
included self-study modules
For future projects
While we analysed the findings, we deemed the following questions worthwhile to investigate in
future:
1.
Were statistical methods and software used in suitable fashion, considering the complexity of the
methods and tools and the implications of conclusions?
2.
Would conclusions be the same for years other than the 2009 sample?
3.
Would a module-based analysis be meaningful and what would such an analysis yield? For
example, are there any modules where the first two assignments are not useful predictors of the
final outcome?
4.
Do courses with traditional high failure rates, perhaps ones such as Accounting and Law, also
display a pattern similar to the overall trend?
5.
BBS is a multi-year multi-specialization degree programme – is the pattern the same at each level
of year of study?
6.
Do international students by any chance progress differently from domestic students (for yet
unidentified reasons)?
7.
Is the pattern for international students under pathway arrangements (i.e. starting here with
second and third year modules) any different?
8.
Would the statistical findings be any different if we only looked at students doing a course for the
first time (i.e. excluding students repeating modules after having failed)?
9.
Is there possibly a core group of students causing most assessment and module failures and how
should they be supported?
10. Do GradDip students with an earlier degree in a business field perform different from other
students?
11. Does the pattern of progression for students change over the study period, especially when doing
the multi-year degree?
12. How does this compare to other qualifications and institutions, including analysis and publication
by Potgieter with various co-authors during 2009 and 2010?
Conclusion
In a general sense, we conclude that the ITP sector could profitably develop statistical models to
predict learner success or failure. To this end, we need to identify those predictors that explain the
most significant proportion of the variance of end-of-course pass and fail categories. This paper, we
reasoned, would be a tentative first step.
Mindful of the limitations of our approach, we soon sensed that the statistical model we were
working on to predict success and failure was questionable on two grounds:
Firstly, of the total cohort of 2160 student data sets, we excluded 370 sets which had missing
values. To develop a more nuanced approach, we would have to look at the reasons for these missing
values, coding the reasons as variables to be included in the data set. Secondly, of those student data
protocols retained in the data set, it seems that more than 80% performed at levels consistent with
their initial two assessment scores. Given (a), we would have to be cautious to interpret initial
assessment scores as consistent with final outcome. Moreover, we need to acknowledge that the
strategy of coding non-academic variables could yield non-training-related predictor variables that
explain a significant proportion of the variance on the pass/fail category. To improve the predictive
efficiency of the model, we would have to re-code the existing data and extend the data set to account
for biographical, motivational and socio-economic variables (to name a few).
Although 17.1% of the total number of data sets have been excluded, we could argue that for
79% of the cohort (2160 - [370 excluded + 84 failures]), their initial assessments were consistent with
final mark. Moreover, of the 1790 student data sets retained in the data set, only 84 have failed. This
reinforces our conclusion that we need to look beyond training-related variables, albeit that the latter
should not be excluded, to develop more nuanced accounts of student performance and attrition.
Indeed, we may discover that our focus should be on factors other than the academic to improve the
effectiveness and efficiency of our interventions.
In summary, we want to argue the case for a comprehensive, nuanced approach to learner
support and pastoral care, looking at learners as complex beings whose functioning in training-related
and non-training-related contexts require individual and customized attention. These decisions, we
argue, have to be informed by reliable and valid information about student worlds of experience and
performance. To return to our initial thesis: we should use prediction models to benefit the training
of all students – our focus on at-risk students is a deficit model. We have to add value to the training
of all students, and good students should become better students, while we also deal effectively and
efficiently with those at-risk. If we have reliable and valid assessment practices in place, we may
develop models to select those students who sit (or do not sit) for final examinations.
References and Citations
Academic Directions 2010. (2010). Faculty Planning document, WINTEC.
Code of Practice for the Pastoral care of International Students. (2003). Ministry of Education,
New Zealand. Accessed 09 March 2009 at
http://www.minedu.govt.nz/educationSectors/InternationalEducation/ForProvidersOfInternationalEd
ucation/CodeofPracticeforInternationalStudents.aspx
First Year Experience (FYE) Retention strategy. (2008) Planning document of PERSON-A,
Manager: Student Experience, WINTEC
NZ Herald. Tertiary education funding to be performance linked. 2010. Retrieved March 15, 2010
from http://www.nzherald.co.nz/nz/news/article.cfm?c_id=1&objectid=10630881
NZ Herald. Time-limit plan for student loan test. 2010. Retrieved March 15, 2010 from
http://www.nzherald.co.nz/nz/news/article.cfm?c_id=1&objectid=10631773
Our Direction. (2009). Presentation by Chief Executive of WINTEC, May 2009.
Potgieter, C., Ferguson, B. (2009) An Analysis of Completion and Retention of a Graduate
Diploma. Proceedings of New Zealand Applied Business Education Conference, Rotorua, 28-29
Sep 2009.
Potgieter, C, Ferguson, B. (2009). Managing International Students Attendance with Consideration
of Completion and Satisfaction. Proceedings of the 22nd conference NACCQ, 10-13 July 2009,
Napier, New Zealand. pp 87-90.
Potgieter, C., Ferguson, B. (2009). Analysis of completion and retention of a graduate diploma.
Presentation at 22nd conference NACCQ, 10-13 July, 2009, Napier, New Zealand. Retrieved
programme September 29, 2009 from
http://hyperdisc.unitec.ac.nz/naccq09/programme/programme_saturday.html.
Potgieter, C., Greyling, W., Ferguson, B. (2010). Managing “At Risk” students and pass rates with
SPSS. Proceedings of the 23rd conference of NACCQ, 6-9 July 2010, Dunedin, New Zealand, pp
227-234.
Potgieter, C., Roberton, G. (2009). Student success: Overcoming quality systems and data
problems. Presentation at 22nd conference NACCQ, 10-13 July, 2009, Napier, New Zealand.
Retrieved September 26, 2009 from
http://hyperdisc.unitec.ac.nz/naccq09/programme/programme_saturday.html.
Student Experience Journey. (2009). Planning document of Jo Shortland, Manager: Student
Experience, WINTEC, 2009
Our Directions. (2009). Staff presentations by Chief Executive, WINTEC.
Scott, D. (2005) Retention, completion and progression in tertiary education in New Zealand, in
Journal of Higher Education Policy and Management, Vol. 27, Issue 1, pp3-17, accessed 15 August
2008 at
http://www.educationcounts.govt.nz/publications/tertiary_education/retention,_completion_and_pro
gression_in_tertiary_education_in_new_zealand
Stepwise Discriminant Analysis. http://marketing.byu.edu/htmlpages/books/pcmds/DISCRIM.html ;
last accessed on 5 July 2010.
Stepwise linear regression in SPSS. http://www.statisticshell.com/multireg.pdf ; last accessed on 5
July 2010.
Tertiary Education Strategy Monitoring Information Website, accessed 15 August 2008.
http://wiki.tertiary.govt.nz/~TESMon/MonitoringInformation/CompletionRetentionAndProgression
Tertiary Education Strategy 2007 - 2012 (2007). Wellington: Ministry of Education, accessed 15
August 2008 at https://staff.wintec.ac.nz/corporatedocs/index.htm
TEC to publish educational performance information. 2010. Accessed 15 March 2010 at
http://www.tec.govt.nz/About-us/News/Media-releases/Media-release-TEC-to-publish-educationalperformance-information/
Tredoux, C., Durheim, K. (2002). Numbers, Hypotheses & Conclusions. A course in statistics for the
social sciences. UCT Press: Cape Town.
Yin, R. (1993). Applications of case study research. Newbury Park, CA: Sage Publishing.
Stake, R. (1995). The art of case research. Newbury Park, CA: Sage Publications
Download