S I R

advertisement
SUMMER INSTITUTE SURVEY REPORT
1 DESCRIPTIVES
Is this your First, Second, or Third time at the Summer Institute?
Count
st
1
2nd
3rd
No Response
30
13
3
6
%
65
28
7
-
Have you Taught any Applied Learning Courses before?
Yes
No
No Response
How many ETEAL Proposals for Funding have you submitted?
0
1
2
3+
No Response
25 76
8 24
19
-
25 58
8 19
7 16
3 7
9
-
How many of your ETEAL Proposals have been funded?
0
1
2
3+
No Response





28 54
8 15
4 8
3 6
9
-
30 Departments represented across the survey population
52 total Survey Respondents
44 Participants completed the Pre-Institute Survey
39 Participants completed the Post-Institute Survey
~75% of participants completed both pre and post surveys
For the majority of participants, this was their very first time at the Summer Institute, though of those
about half were accompanied by colleagues from the same department who had been once or even
twice before. Most participants who attended had taught an Applied Learning course before, though
this question had a high rate of non-response. This could be because there were many staff members
who did not teach at all present on the first day, but it could also be because of a typo on the survey
itself which made the question read “I have taught _____ Applied Learning” rather than “I have taught
_____ Applied Learning Courses” This also led to confusion among respondents about whether they
should be checking the blank as a yes/no question or filling in a specific number.
While around half of all respondents reported having taught Applied Learning courses in the
past, just as many respondents reported never having submitted a proposal for ETEAL funding. From the
data we collected, it seems that among Summer Institute attendees, the chances of being funded if you
submit an ETEAL proposal are fairly high. This is generally similar to the overall rate of ETEAL funding and
the rates of funding over the past year, though it is still on the higher end of that range. In general, since
fall of 2013 ETEAL proposals have been funded at a rate of about two out of every three, while the rate
of funding was just over 70% among survey respondents.
This in itself is not terribly significant, but when we look closer we can see that only three
attendees had submitted proposals but never received funding. The rest of the respondents had all
received funding at least once, even if they were rejected on other occasions. Put more plainly, only
three respondents who had submitted proposals but never received funding were in attendance. As we
will see in the next section, the Summer Institute appeared to have a significant positive impact on how
confident respondents felt in their ability to submit a worthwhile proposal for funding, so it would
benefit those who have previously not received funding the most.
2 PRE AND POST TESTING
Change in Attitudes, Opinions, and Plans between day 1 and day 2
n
Initial Survey
Mean
4.15
Exit Survey
Mean
4.30
Difference
in Means
0.15
I feel that I know how to design an Applied Learning 30
Experience
4.06
4.48
0.42
I have a clear understanding of how applied learning 32
impacts students and instructors
3.45
4.45
1.00
I feel that I know exactly what I need to do to submit 30
a proposal for ETEAL funding
32
4.73
4.91
0.19
To what Extent do you think Applied Learning
Experiences can enhance student learning?
3.78
3.97
0.18
How Well do you feel you could Implement an ETEAL 28
Applied Learning Experience in one of your courses?
31
4.78
4.84
0.07
I feel that Applied Learning Experiences provide
Students with Valuable Skills and Experience
3.69
4.19
0.5
I have all of the Resources and Information I need to 28
Carry out an Applied Learning Experience within the
next year
28
4.65
4.94
0.29
How Likely do you think you are to plan an Applied
Learning Experience within the next year?
Difference in Means calculated using a Paired T-test, cases with missing pair data excluded
“I don’t know” coded as missing data, cases not reflected here, see discussion below
1
Significant at the 0.01 level
2 Significant at the 0.1 level, not substantively significant
Is it
Significant?
No
Yes1
Yes1
Yes2
No
No
No
No
Right away, we can see that some opinions and attitudes changed more than others over the
course of the two days of the Summer Institute. The differences in means column lists the results of a
paired t-test used to determine the overall average difference between the pre-survey and post-survey
answers for each respondent. This tells us how much their answers changed over the course of the
Institute, as well as which direction that change was in. By far the most significant change, both
substantively and statistically, is the shift in confidence regarding the submission of ETEAL funding
proposals. Most respondents reported feeling like they knew substantially more about how to submit an
ETEAL proposal by the end of the Institute. Among all the survey questions, this item showed the biggest
overall increase between the pre and post surveys, meaning that overall, respondents felt like they
gained a noticeable amount of knowledge and/or confidence regarding the submission of ETEAL
proposals.
The next item of note is the seventh question, regarding a respondent’s feeling that they have
the information and resources needed to carry out an Applied Learning Experience within the next
academic year. The change here was not as drastic as it was in the case of the ETEAL proposals, but by
the end of the Institute respondents did, on average, appear to feel more confident that they had the
resources and understanding needed to engage students in Applied Learning. This is, however, contrary
to the overall lower average confidence attendees had in their ability to implement an ETEAL funded
initiative in one of their courses. The change here was not statistically significant; in fact, there was
almost no change at all, and this is perhaps a cause for concern. By the end of the Institute, respondents
felt no more confident in their ability to implement an ETEAL funded initiative.
Respondent confidence in their ability to design Applied Learning Experiences rose over the
course of the institute, as did their perceived understanding of how Applied Learning impacted students
and instructors and how Applied Learning could enhance student learning. When we look more closely
at the responses for these items, we can also see that while most of the increase is the result of
respondents who initially answered “Disagree” or “Neither Agree nor Disagree” changing their answers
to either “Agree” or “Strongly Agree” in the exit survey, there were also a small number of respondents
whose opinions decreased by the end of the Institute. Although there were not many and it is difficult to
make any definitive claims about what it would mean given such a small sample size, there were two
respondents who initially felt that they knew how to design Applied Learning Experiences at the start of
the Institute and then by the end were not that certain. Given the impact that every instructor has on
their students, we should pay attention to these cases, few though they are, to make sure that the
Summer Institute does not have a negative impact on any of its participants.
Overall, it should be noted that the net effect of the Summer Institute appears to be fairly
positive on all fronts, with most respondents showing slight gains in their opinion of Applied Learning,
their confidence in implementing it, and their perceived knowledge about Applied Learning and
submitting ETEAL proposals. Although there was a roughly 75% response rate overall, there were seven
respondents who never completed their exit surveys and six who never completed initial surveys,
combined with those who left individual questions incomplete this leaves us with thirteen respondents
whose experiences are not accurately represented here.
This brings us to the issue of Significance. Several items were rated as statistically significant,
two at the 0.01 level and one at the 0.1 level, but this may be misleading. Substantively, there are only
three questions in which attendees appeared to have relevant and noteworthy gains. If we look at the
actual beginning and ending means, we can see that for several items such as question 4 (regarding the
extent to which attendees felt that Applied Learning enhanced student learning) that on average
respondents either marked “Agree” or “Strongly Agree” and by the end of the Institute, there was
simply a slight shift toward “Strongly Agree”. We can argue then that this is not substantively significant
because the shift between “Agree” and “Strongly Agree” is not likely to be a dramatic or noticeable
change. There were also a number of respondents who answered items with “I don’t Know” or “I’m not
sure” during their initial surveys and these cases were excluded from the overall paired-t tests, but they
are still important cases. Both Attendees who answered that they Didn’t Know how they felt about
Applied Learning design at the start of the Institute changed their answers to “Agree” or “Strongly
Agree” by the end. For most all attendees, this was the case no matter what the question and in all but
one case, attendees moved from “I don’t know” to “agree” or an equivalent answer. One attendee
maintained their “I don’t know” answer throughout the institute, and two others who answered “I don’t
know” did not complete their exit surveys.
Finally, we should note that while there was clearly a noticeable and significant change in
Attendee attitudes and opinions by the end of the Institute, we cannot make any claims of causation.
Not only are we limited by the relatively small sample size and the low rate of survey completion, these
surveys could not accurately capture any information about which part of the Summer Institute would
have affected these changes. In fact, we cannot be sure that these changes are the result of the Institute
at all, only that they occurred between the time when the Institute began and when it ended. Still, this
provides us with valuable insight into the areas that attendees improved in the most, as well as letting
us know where they already felt confident and competent.
3 EXPECTATIONS AND EXPERIENCES
The final question on both the initial and exit surveys was an open ended item designed to be a sort
of reflective piece, a parallel to the intention and critical reflection prompts used in Applied
Learning in ETEAL funded courses. At the start of the Summer Institute, respondents were asked
what their goals were for the institute. At the end, we asked them what they felt they gained by
attending. Compared to the body of the survey, the response rate on this final question was much
lower and although almost everyone answered at least one of the two reflections, only 21
respondents answered both. There are several possible reasons for this, though the primary culprit
is likely time considering that it takes much longer to think about and write a reflective response
than to check off boxes. Still, despite the low response rate, several distinct patterns emerged that
can tell us a lot about the expectations and experiences of the Summer Institute Attendees.
Among all of the participants who expressed clear expectations for the Summer Institute there
were four common themes but no single expectation represented the majority. The most common
theme by far was a desire to gain information or learn more about Applied Learning or how to
implement it in a course, which was the goal of at least 19 attendees. Many (11) respondents also
indicated a desire for new ideas and information in general which may or may not have been
synonymous with the previous point, based on their understanding of the framing of the Summer
Institute. Following that, 8 respondents were concerned with learning more about AL concepts and
theories themselves as opposed to just ideas for courses, 7 were focused on networking and
collaborative opportunities, and 6 were looking for support or information about things that supported
AL such as critical reflection and measuring student learning outcomes. Other goals and expectations
included learning about using Applied Learning in online courses, learning more about ETEAL funding
proposals, and generally improving teaching skills.
Overall, expectations were largely centered on the exchange of ideas, information, and
developing a knowledge base that would help individual instructors with their own courses. After that,
we start to see themes of cooperation and collaboration, but when compared with all of the
expectations that focus more on an individualized teaching goal or practice, the desire for collaboration
and networking is relatively small.
Breaking away from the more individually focused goals indicated on the initial survey, more
participants reported networking, collaboration, and the exchange of ideas as their primary gain from
the Summer Institute. This is not only a valuable effect and insight but also a source of even more
important questions; if the Summer Institute has the capacity to build camaraderie and increase
interdisciplinary communication and cooperation, then which part of the Institute is responsible? Was it
the round table discussions? Was it the Speaker-facilitated activities? Was it just lunch? If we can find
out which aspects of the Institute were most effective for networking and for developing
interdepartmental communication, then perhaps we could distill that and find a way to have other,
similar one-day events that focus specifically on interdisciplinary networking within the framework of an
Applied Learning marketplace of ideas and skills. Also, with any event that builds interdisciplinary
bridges, we’d need to establish some form of at least short-term monitoring to see how much
interdisciplinary communication and collaboration there is after these events. It’s entirely possible that
the exchange of ideas and interdisciplinary communication are only really happening in these events
and not outside of them.
Several participants also had specific comments about the guest speakers at the Summer
Institute. Peg Wessinger, the speaker on Day one, received the most praise from participants and her
topics and activities were well received among those who spoke of her, though it should be noted that
only six participants mentioned her at all. There was an equal mix of positive and negative commentary
on Laura Cruz, the speaker of the second day. It’s worth noting that the criticisms she received were
fairly targeted while the positive comments she received were often indirect or lumped in with general
positive statements such as “I enjoyed all of the guest speakers”. Because of the small number of
participants who commented on Laura Cruz, however, it’s hard to know the overall impression she left
on all participants.
Among those who responded to the final question on both the entry and exit survey, 15
reported achieving the goals they had at the start of the institute or even achieving something
additional. 5 Did not mention their original goals but instead reported that they had gained something
entirely different than what they set out to find. While this hints at a high level of satisfaction and
efficacy, we must be careful about making such a declaration. Not all participants who attended the
institute completed an exit survey and among the participants who did, 12 declined to provide a final
statement about what they felt they gained by attending the Summer Institute.
4 THE EFFECT OF ATTENDANCE AND RECOMMENDATIONS
From the results of the initial and exit surveys, we can see that the Summer Institute clearly had
an effect on those who attended and in a majority of cases this effect was entirely positive. With few
exceptions, every respondent reported feeling either better or the same about their understanding of
Applied Learning, its impact on students and instructors, its design and implementation, and its overall
benefit. Attendees also rated their understanding of how to submit an ETEAL proposal much higher by
the end of the Institute and generally felt that they had a better grasp of the resources and information
they’d need to carry out an Applied Learning Experience within the next year. Although the Institute was
not significantly likely to change an attendee’s plans to design a new Applied Learning Experience, most
attendees were already planning on that at the start of the Institute.
The exchange of ideas and the collaborative atmosphere of the Institute was also reported as a
common experience and benefit, though subsequent monitoring will be necessary to determine how
resilient the connections formed at the Summer Institute are. Regardless of longevity though, the
Summer Institute produced a distinct feeling of collaboration and camaraderie among attendees which
has value in itself. We can only speculate about which aspect of the Institute is responsible for this sense
of collaboration and mutual exchange at present, but upcoming follow-up interviews with attendees
should clarify the matter. It should also be noted that there were a large number of academic
departments not represented at the Institute and so there were perspectives that attendees did not
have direct access to. In the future, increasing both the representation of each academic department
and facilitating further collaboration and contact could benefit the Applied Learning Teaching
Community overall.
One avenue for potential improvement of the Summer Institute lies in continuing to refine it as
an applied learning exercise. The Institute already has most of the components of an Applied Learning
Experience, especially when it comes to the critical discussions that faculty have with one another and
the speakers, but it is still possible to make the Institute even more applied. The final question on both
the initial and exit surveys, for example, could be expanded to serve as Intention and Critical reflections.
An applied learning course creation activity could be added for the whole group with group discussion,
critique, and speaker feedback built in to further integrate the applied practice of designing an Applied
Learning Experience.
Another option for improvement, based on the results we obtained, would be to allow previous
ETEAL funded instructors to introduce themselves to the group at the start of the institute so that they
could be identified as potential collaborators or mentors by those looking for advice. There was such an
introduction in this institute, but it came at the end of the second day after many people were either
preparing to leave or had already gone. If networking and forming collaborative partnerships was one of
the greatest gains that attendees felt they made, then this could help them spend their time networking
more efficiently. Ideally, introductions for every Attendee would maximize this benefit but that is not
likely to be feasible given the size of the Summer Institute. It may also be beneficial to nurture the
connections made at the Summer Institute throughout the fall semester by holding shorter events
focused specifically on faculty and staff networking and collaboration.
As a final note and recommendation, the response rate on future Summer Institute surveys
might be increased by translating the paper survey to an electronic form that respondents could answer
via clicker. This would serve the dual purpose of gather data and introducing attendees to technology
that they could use in their own classrooms.
Download