Uploaded by Bleasy Clyde Rodel

Thematic analysis

advertisement
AISHE-J
Volume , Number 3 (Autumn 2017)
3351
Doing a Thematic Analysis: A Practical, Step-by-Step
Guide for Learning and Teaching Scholars.
*
Moira Maguire & Brid Delahunt
Dundalk Institute of Technology.
Abstract
Data analysis is central to credible qualitative research. Indeed the qualitative researcher is
often described as the research instrument insofar as his or her ability to understand, describe
and interpret experiences and perceptions is key to uncovering meaning in particular
circumstances and contexts. While much has been written about qualitative analysis from a
theoretical perspective we noticed that often novice, and even more experienced researchers,
grapple with the ‘how’ of qualitative analysis. Here we draw on Braun and Clarke’s (2006)
framework and apply it in a systematic manner to describe and explain the process of analysis
within the context of learning and teaching research. We illustrate the process using a worked
example based on (with permission) a short extract from a focus group interview, conducted
with undergraduate students.
Key words: Thematic analysis, qualitative methods.
Acknowledgements.
We gratefully acknowledge the support of National Digital Learning Repository (NDLR) local
funding at DkIT in the initial development of this work.
*URL: http://ojs.aishe.org/index.php/aishe-j/article/view/335
AISHE-J
Volume 8, Number 3 (Autumn 2017)
3352
1. Background.
Qualitative methods are widely used in learning and teaching research and scholarship
(Divan, Ludwig, Matthews, Motley & Tomlienovic-Berube, 2017). While the epistemologies
and theoretical assumptions can be unfamiliar and sometimes challenging to those from, for
example, science and engineering backgrounds (Rowland & Myatt, 2014), there is wide
appreciation of the value of these methods (e.g. Rosenthal, 2016). There are many, often
excellent, texts and resources on qualitative approaches, however these tend to focus on
assumptions, design and data collection rather than the analysis process per se.
More and more it is recognised that clear guidance is needed on the practical aspects of how
to do qualitative analysis (Clarke & Braun, 2013). As Nowell, Norris, White and Moules (2017)
explain, the lack of focus on rigorous and relevant thematic analysis has implications in terms
of the credibility of the research process. This article offers a practical guide to doing a
thematic analysis using a worked example drawn from learning and teaching research. It is
based on a resource we developed to meet the needs of our own students and we have used
it successfully for a number of years. It was initially developed with local funding from[Irish]
National Digital Learning Repository (NDLR) and then shared via the NDLR until this closed in
2014. In response to subsequent requests for access to it we decided to revise and develop
this as an article focused more specifically on the learning and teaching context. Following
Clarke & Braun’s (2013) recommendations, we use relevant primary data, include a worked
example and refer readers to examples of good practice.
2. Thematic Analysis.
Thematic analysis is the process of identifying patterns or themes within qualitative data.
Braun & Clarke (2006) suggest that it is the first qualitative method that should be learned as
‘..it provides core skills that will be useful for conducting many other kinds of analysis’ (p.78).
A further advantage, particularly from the perspective of learning and teaching, is that it is a
method rather than a methodology (Braun & Clarke 2006; Clarke & Braun, 2013). This means
that, unlike many qualitative methodologies, it is not tied to a particular epistemological or
theoretical perspective. This makes it a very flexible method, a considerable advantage given
the diversity of work in learning and teaching.
AISHE-J
Volume 8, Number 3 (Autumn 2017)
3353
There are many different ways to approach thematic analysis (e.g. Alhojailan, 2012;
Boyatzis,1998; Javadi & Zarea, 2016). However, this variety means there is also some
confusion about the nature of thematic analysis, including how it is distinct from a qualitative
content analysis1 (Vaismoradi, Turunen & Bonda, 2013). In this example, we follow Braun &
Clarke’s (2006) 6-step framework. This is arguably the most influential approach, in the social
sciences at least, probably because it offers such a clear and usable framework for doing
thematic analysis.
The goal of a thematic analysis is to identify themes, i.e. patterns in the data that are important
or interesting, and use these themes to address the research or say something about an
issue. This is much more than simply summarising the data; a good thematic analysis
interprets and makes sense of it. A common pitfall is to use the main interview questions as
the themes (Clarke & Braun, 2013). Typically, this reflects the fact that the data have been
summarised and organised, rather than analysed.
Braun & Clarke (2006) distinguish between two levels of themes: semantic and latent.
Semantic themes ‘…within the explicit or surface meanings of the data and the analyst is not
looking for anything beyond what a participant has said or what has been written.’ (p.84). The
analysis in this worked example identifies themes at the semantic level and is representative
of much learning and teaching work. We hope you can see that analysis moves beyond
describing what is said to focus on interpreting and explaining it. In contrast, the latent level
looks beyond what has been said and ‘…starts to identify or examine the underlying ideas,
assumptions, and conceptualisations – and ideologies - that are theorised as shaping or
informing the semantic content of the data’ (p.84).
3. The Research Question And The Data.
The data used in this example is an extract from one of a series of 8 focus groups involving 40
undergraduate student volunteers. The full study involved 8 focus-groups lasting about 40
minutes. These were then transcribed verbatim. The research explored the ways in which
students make sense of and use feedback. Discussions focused on what students thought
about the feedback they had received over the course of their studies: how they understood it;
the extent to which they engaged with it and if and how they used it. The study was ethically
approved by the Dundalk Institute of Technology School of Health and Science Ethics
Committee. All of those who participated in the focus group from which the extract is taken
1 See O’Cathain & Thomas (2004) for a useful guide to using content analysis on responses to openended survey questions.
AISHE-J
Volume 8, Number 3 (Autumn 2017)
3354
also gave permission for the transcript extract to be used in this way.
The original research questions were realist ones – we were interested in students’ own
accounts of their experiences and points of view. This of course determined the interview
questions and management as well the analysis. Braun & Clarke (2006) distinguish between a
top-down or theoretical thematic analysis, that is driven by the specific research question(s)
and/or the analyst’s focus, and a bottom-up or inductive one that is more driven by the data
itself. Our analysis was driven by the research question and was more top-down than bottom
up. The worked example given is based on an extract (approx. 15 mins) from a single focus
group interview. Obviously this is a very limited data corpus so the analysis shown here is
necessarily quite basic and limited. Where appropriate we do make reference to our full
analysis however our aim was to create a clear and straightforward example that can be used
as an accessible guide to analysing qualitative data.
3.1 Getting started.
The extract: This is taken from a real focus-group (group-interview) that was conducted with
students as part of a study that explored student perspectives on academic feedback. The
extract covers about 15 minutes of the interview and is available in Appendix 1.
Research question: For the purposes of this exercise we will be working with a very broad,
straightforward research question: What are students’ perceptions of feedback?
3.2 Doing the analysis.
Braun & Clarke (2006) provide a six-phase guide which is a very useful framework for
conducting this kind of analysis (see Table 1). We recommend that you read this paper in
conjunction with our worked example. In our short example we move from one step to the
next, however, the phases are not necessarily linear. You may move forward and back
between them, perhaps many times, particularly if dealing with a lot of complex data.
Step 1: Become familiar with the data,
Step 4: Review themes,
Step 2: Generate initial codes,
Step 5: Define themes,
Step 3: Search for themes,
Step 6: Write-up.
Table 1: Braun & Clarke’s six-phase framework for doing a thematic analysis
AISHE-J
Volume 8, Number 3 (Autumn 2017)
3355
3.3 Step 1: Become familiar with the data.
The first step in any qualitative analysis is reading, and re-reading the transcripts. The
interview extract that forms this example can be found in Appendix 1.
You should be very familiar with your entire body of data or data corpus (i.e. all the interviews
and any other data you may be using) before you go any further. At this stage, it is useful to
make notes and jot down early impressions. Below are some early, rough notes made on the
extract:
The students do seem to think that feedback is important but don’t always find it useful.
There’s a sense that the whole assessment process, including feedback, can be seen as
threatening and is not always understood. The students are very clear that they want very
specific feedback that tells them how to improve in a personalised way. They want to be able
to discuss their work on a one-to-one basis with lecturers, as this is more personal and also
private. The emotional impact of feedback is important.
3.4 Step 2: Generate initial codes.
In this phase we start to organise our data in a meaningful and systematic way. Coding
reduces lots of data into small chunks of meaning. There are different ways to code and the
method will be determined by your perspective and research questions.
We were concerned with addressing specific research questions and analysed the data with
this in mind – so this was a theoretical thematic analysis rather than an inductive one. Given
this, we coded each segment of data that was relevant to or captured something interesting
about our research question. We did not code every piece of text. However, if we had been
doing a more inductive analysis we might have used line-by-line coding to code every single
line. We used open coding; that means we did not have pre-set codes, but developed and
modified the codes as we worked through the coding process.
We had initial ideas about codes when we finished Step 1. For example, wanting to discuss
feedback on a one-to one basis with tutors was an issue that kept coming up (in all the
interviews, not just this extract) and was very relevant to our research question. We discussed
these and developed some preliminary ideas about codes. Then each of us set about coding
a transcript separately. We worked through each transcript coding every segment of text that
seemed to be relevant to or specifically address our research question. When we finished we
compared our codes, discussed them and modified them before moving on to the rest of the
transcripts. As we worked through them we generated new codes and sometimes modified
AISHE-J
Volume 8, Number 3 (Autumn 2017)
3356
existing ones. We did this by hand initially, working through hardcopies of the transcripts with
pens and highlighters. Qualitative data analytic software (e.g. ATLAS, Nvivo etc.), if you have
access to it, can be very useful, particularly with large data sets. Other tools can be effective
also; for example, Bree & Gallagher (2016) explain how to use Microsoft Excel to code and
help identify themes. While it is very useful to have two (or more) people working on the
coding it is not essential. In Appendix 2 you will find the extract with our codes in the margins.
3.5 Step 3: Search for themes.
As defined earlier, a theme is a pattern that captures something significant or interesting about
the data and/or research question. As Braun & Clarke (2006) explain, there are no hard and
fast rules about what makes a theme. A theme is characterised by its significance. If you have
a very small data set (e.g. one short focus-group) there may be considerable overlap between
the coding stage and this stage of identifying preliminary themes.
In this case we examined the codes and some of them clearly fitted together into a theme. For
example, we had several codes that related to perceptions of good practice and what students
wanted from feedback. We collated these into an initial theme called The purpose of
feedback.
At the end of this step the codes had been organised into broader themes that seemed to say
something specific about this research question. Our themes were predominately descriptive,
i.e. they described patterns in the data relevant to the research question. Table 2 shows all
the preliminary themes that are identified in Extract 1, along with the codes that are associated
with them. Most codes are associated with one theme although some, are associated with
more than one (these are highlighted in Table 2). In this example, all of the codes fit into one
or more themes but this is not always the case and you might use a ‘miscellaneous’ theme to
manage these codes at this point.
AISHE-J
Volume 8, Number 3 (Autumn 2017)
Theme : The purpose of Theme: Lecturers.
feedback.
Codes
Codes
Ask some Ls,
Help to learn what you’re doing
Some Ls more approachable,
wrong,
3357
Theme: Reasons for using feedback (or not).
Codes
To improve grade,
Limited feedback,
Didn’t understand fdbk,
U n a b l e t o j u d g e w h e t h e r Some Ls give better advice,
question has been answered,
Reluctance to admit difficulties to L,Fear Fdbk focused on grade ,
of unspecified disadvantage,
Use to improve grade,
U n a b l e t o j u d g e w h e t h e r Unlikely to approach L to discuss fdbk,
Distinguish purpose and use,
question interpreted
Lecturer variability in framing fdbk,
Unlikely to approach L to discuss fdbk,
properly,
Unlikely to make a repeated attempt,
Improving structure improves grade,
Distinguish purpose and use,
Have discussed with tutor,
Can’t separate grade and learning,
Improving grade,
Example: Wrong frame of mind
New priorities take precedence = forget about
Improving structure
feedback
Theme: How feedback is used T h e m e : E m o t i o n a l r e s p o n s e t o Theme: What students want from feedback.
(or not).
feedback.
Codes
Codes
Codes
Usable fdbk explains grade and how to improve,
Read fdbk,
Like to get fdbk,
Want fdbk to explain grade,
Usually read fdbk,
Don’t want to get fdbk if haven’t done
Example- uninformative fdbk,
well,
Refer to fdbk if doing
Very specific guidance wanted,
Reluctance to hear criticism,
same subject,
Reluctance to hear criticism (even if More fdbk wanted,
Not sure fdbk is used,
constructive),
Want dialogue with L,
Used fdbk to improve
Fear of possible criticism,
Dialogue means more,
referencing,
Experience: unrealistic fear of criticism,
Dialogue more personalised/ individual,
Example: using fdbk to
Fdbk taken personally initially,
Dialogue more time consuming but better,
improve referencing,
Fdbk has an emotional impact,
Want dedicated class for grades and fdbk,
Refer back to example
Difficult for L to predict impact,
Compulsory fdbk class,
that ‘went right’,
Student variability in response to fdbk,
Structured option to get fdbk,
Forget about fdbk until
Want fdbk in L’s office as emotional
Fdbk should be constructive,
response difficult to manage in public,
next assignment,
Fdbk should be about the work and not the
Wording doesn’t make much difference,
Fdbk applicable to similar
person,
Lecturer variability in framing fdbk,
assignments,
Experience – fdbk is about the work,
Negative fdbk can be constructive,
Fdbk on referencing
Difficulties judging own work,
Negative fdbk can be framed in a
widely applicable,
Want fdbk to explain what went right,
supportive way.
Experience: fdbk focused
Fdbk should focus on understanding,
on referencing,
Improving understanding improves grade.
Generic fdbk widely
Want fdbk in Ls office as emotional response
difficult to manage in public.
applicable.
Table 2: Preliminary themes (* fdbk = feedback; L = lecturers)
AISHE-J
Volume 8, Number 3 (Autumn 2017)
3358
3.6 Step 4: Review themes.
During this phase we review, modify and develop the preliminary themes that we identified in
Step 3. Do they make sense? At this point it is useful to gather together all the data that is
relevant to each theme. You can easily do this using the ‘cut and paste’ function in any word
processing package, by taking a scissors to your transcripts or using something like Microsoft
Excel (see Bree & Gallagher, 2016). Again, access to qualitative data analysis software can
make this process much quicker and easier, but it is not essential. Appendix 3 shows how the
data associated with each theme was identified in our worked example. The data associated
with each theme is colour-coded.
We read the data associated with each theme and considered whether the data really did
support it. The next step is to think about whether the themes work in the context of the entire
data set. In this example, the data set is one extract but usually you will have more than this
and will have to consider how the themes work both within a single interview and across all
the interviews.
Themes should be coherent and they should be distinct from each other. Things to think about
include:
•
Do the themes make sense?
•
Does the data support the themes?
•
Am I trying to fit too much into a theme?
•
If themes overlap, are they really separate themes?
•
Are there themes within themes (subthemes)?
•
Are there other themes within the data?
For example, we felt that the preliminary theme, Purpose of Feedback ,did not really work as a
theme in this example. There is not much data to support it and it overlaps with Reasons for
using feedback(or not) considerably. Some of the codes included here (‘Unable to judge
whether question has been answered/interpreted properly’) seem to relate to a separate issue
of student understanding of academic expectations and assessment criteria.
We felt that the Lecturers theme did not really work. This related to perceptions of lecturers
and interactions with them and we felt that it captured an aspect of the academic environment.
We created a new theme Academic Environment that had two subthemes: Understanding
AISHE-J
Volume 8, Number 3 (Autumn 2017)
3359
Academic Expectations and Perceptions of Lecturers. To us, this seemed to better capture
what our participants were saying in this extract. See if you agree.
The themes, Reasons for using feedback (or not), and How is feedback used (or not) ,did not
seem to be distinct enough (on the basis of the limited data here) to be considered two
separate themes. Rather we felt they reflected different aspects of using feedback. We
combined these into a new theme Use of feedback, with two subthemes, Why? and How?
Again, see what you think.
When we reviewed the theme Emotional Response to Feedback we felt that there was at least
1 distinct sub-theme within this. Many of the codes related to perceptions of feedback as a
potential threat, particularly to self-esteem and we felt that this did capture something
important about the data. It is interesting that while the students’ own experiences were quite
positive the perception of feedback as potentially threatening remained.
So, to summarise, we made a number of changes at this stage:
•
We eliminated the Purpose of Feedback theme,
•
We created a new theme Academic Environment that had two subthemes:
Understanding Academic Expectations and Perceptions of Lecturers,
•
We collapsed Purpose of Feedback, Why feedback is (not)used and How feedback is
(not) used into a new theme, Use of feedback,
•
We identified Feedback as potentially threatening as a subtheme within the broader
theme Emotional Response to feedback.
These changes are shown in Table 3 below. It is also important to look at the themes with
respect to the entire data set. As we are just using a single extract for illustration we have not
considered this here, but see Braun & Clarke (2006, p 91-92) for further detail. Depending on
your research question, you might also be interested in the prevalence of themes, i.e. how
often they occur. Braun & Clarke (2006) discuss different ways in which this can be addressed
(p.82-82).
AISHE-J
Volume 8, Number 3 (Autumn 2017)
T h e m e : A c a d e m i c Theme: Use of feedback.
Context.
Subtheme: Reasons for using
Subtheme: Academic fdbk (or not).
expectations.
Help to learn what you’re doing
Unable to judge whether wrong,
question has been
Improving grade Improving
answered,
structure,
Unable to judge whether
q u e s t i o n i n t e r p r e t e d To improve grade,
properly,
Limited feedback,
Difficulties judging own
Didn’t understand fdbk,
work.
Fdbk focused on grade,
Subtheme: Perceptions
of lecturers ,
Use to improve grade,
Ask some Ls,
Distinguish purpose and use,
Some
Ls
approachable,
T h e m e : E m o t i o n a l Theme: What students want
response to feedback.
from feedback.
Like to get fdbk,
Usable fdbk explains grade and
how to improve,
Difficult for L to predict
impact,
Example- uninformative fdbk,
Very specific guidance wanted,
Student variability in
response to fdbk,
More fdbk wanted,
S u b t h e m e : F e e d b a c k Want dialogue with L,
potentially threatening.
Dialogue means more,
Don’t want to get fdbk if
Dialogue more personalised/
haven’t done well,
individual,
Reluctance to hear criticism,
Dialogue more time consuming
Reluctance to hear criticism but better,
(even if constructive),
Want dedicated class for
Fear of possible criticism,
grades and fdbk,
m o r e Improving structure improves
grade,
Experience: fear of potential
S o m e L s g i v e b e t t e r C a n ’ t s e p a r a t e g r a d e a n d criticism,
advice,
learning,
Fdbk taken personally
R e l u c t a n c e t o a d m i t New priorities take precedence = initially,
difficulties to L,
forget about feedback.
Fdbk has an emotional
F e a r o f u n s p e c i f i e d Subtheme: How fdbk is used impact,
disadvantage,
(or not).
Want fdbk in L’s office as
emotional response difficult
Unlikely to approach L to Read fdbk/Usually read fdbk,
to manage in public,
discuss fdbk,
Refer to fdbk if doing same
Wording doesn’t make much
U n l i k e l y t o m a k e a subject,
difference,
repeated attempt,
Not sure fdbk is used,
33510
Compulsory fdbk class,
Structured option to get fdbk,
Fdbk should be constructive ,
Fdbk should be about the work
and not the person,
Experience – fdbk is about the
work,
Want fdbk to explain grade,
Want fdbk to explain what went
right,
Negative fdbk can be
Fdbk should focus on
Used fdbk to improve referencing, constructive,
understanding,
Example: Wrong frame of
Negative
fdbk
can
be
framed
mind,
Example: using fdbk to improve
Improving understanding
in a supportive way.
referencing,
improves grade,
Lecturer variability in
framing fdbk.
Refer back to example that ‘went
Want fdbk in L’s office
as
right’,
emotional response difficult to
manage in public.
Forget about fdbk until next
assignment,
Have discussed with tutor,
Fdbk applicable to similar
assignments,
Fdbk on referencing widely
applicable,
Experience: fdbk focused on
referencing,
Generic fdbk widely applicable.
Table 3: Themes at end of Step 4
AISHE-J
Volume 8, Number 3 (Autumn 2017)
33511
3.7 Step 5: Define themes.
This is the final refinement of the themes and the aim is to ‘..identify the ‘essence’ of what
each theme is about.’.(Braun & Clarke, 2006, p.92). What is the theme saying? If there are
subthemes, how do they interact and relate to the main theme? How do the themes relate to
each other? In this analysis, What students want from feedback is an overarching theme that
is rooted in the other themes. Figure 1 is a final thematic map that illustrates the relationships
between themes and we have included the narrative for What students want from feedback
below.
Emotional response
What students want
from feedback
Potential threat
Academic Environment
Perceptions of Ls
Understanding expectations
Use of feedback
Why?
How?
Figure 1: Thematic map.
What students want from feedback.
Students are clear and consistent about what constitutes effective feedback and made
concrete suggestions about how current practices could be improved. What students want
from feedback is rooted in the challenges; understanding assessment criteria, judging their
own work, needing more specific guidance and perceiving feedback as potentially threatening.
Students want feedback that both explains their grades and offers very specific guidance on
how to improve their work. They conceptualised these as inextricably linked as they felt that
improving understanding would have a positive impact on grades.
Students identified that
they not only had difficulties in judging their own work but also how or why the grade was
awarded. They wanted feedback that would help them to evaluate their own work.
AISHE-J
Volume 8, Number 3 (Autumn 2017)
33512
‘Actually if you had to tell me how I got a 60 or 67, how I got that grade, because I
know every time I'm due to get my result for an assignment, I kind of go ‘oh I did so
bad, I was expecting to get maybe 40 or 50’, and then you go in and you get in the
high 60s or 70s. It's like how did I get that? What am I doing right in this piece of work?’
(F1, lines 669-672).
Participants felt that they needed specific, concrete suggestions for improvement that they
could use in future work. They acknowledged that they received useful feedback on
referencing but that other feedback was not always specific enough to be usable.
‘The referencing thing I’ve tried to, that’s the only… that’s really the only feedback we
have gotten back ,I have tried to improve, but everything else it’s just kind of been ‘well
done’, I don’t… hasn’t really told us much.’ (F1, lines 389-392).
Significantly, it emerged that students want opportunities for both verbal and written feedback
from lecturers. The main reason identified for wanting more formal verbal feedback is that it
facilitates dialogue on issues that may be difficult to capture on paper. Moreover, it seems that
feedback enables more specific comments on strengths and limitations of submitted work.
However, it is also clear that verbal feedback is valued as the perception that lecturers are
taking an interest in individual students is perceived to ‘mean more’.
‘I think also the thing that, you know… the fact that someone has sat down and taken
the time to actually tell you this is probably, it gives you an incentive to do it (overspeaking). It does mean a bit more ‘ (M1, lines 456-458).
For these participants, the ideal situation was to receive feedback on a one-to-one basis in the
lecturer’s office.
Privacy is seen as important as students do find feedback potentially
threatening and are concerned about managing their reactions in public. For these students, it
was difficult to proactively access feedback, largely because the demands of new work limited
their capacity to focus on completed work. Given this, they wanted feedback sessions to be
formally scheduled.
3.8 Step 6: Writing-up.
Usually the end-point of research is some kind of report, often a journal article or dissertation.
Table 4 includes a range of examples of articles, broadly in the area of learning and teaching,
that we feel do a good job of reporting a thematic analysis.
Table 4: Some examples of articles reporting thematic analysis.
AISHE-J
Volume 8, Number 3 (Autumn 2017)
33513
Gagnon, L.L. & Roberge, G. (2012). Dissecting the journey: Nursing student experiences
with collaboration during the group work process. Nurse Education Today, 32(8), 945-950.
Karlsen, M-M. W., Wallander; Gabrielsen, A.K., Falch, A.L. & Stubberud, D.G. (2017).
Intensive care nursing students’ perceptions of simulation for learning confirming
communication skills: A descriptive qualitative study. Intensive & Critical Care Nursing, 42,
97-104.
Lehtomäki, E., Moate, J. & Posti-Ahokas, H. (2016). Global connectedness in higher
education: student voices on the value of crosscultural learning dialogue. Studies in Higher
Education, 41 (11), 2011-2027.
Polous, A. & Mahony, M-J. (2008). Effectiveness of feedback: the students' perspectives.
Assessment & Evaluation in Higher Education, 33(2), 143-154.
4. Concluding Comments.
Analysing qualitative data can present challenges, not least for inexperienced researchers. In
order to make explicit the ‘how’ of analysis, we applied Braun and Clarke (2006) thematic
analysis framework to data drawn from learning and teaching research. We hope this has
helped to illustrate the work involved in getting from transcript(s) to themes. We hope that you
find their guidance as useful as we continue to do when conducting our own research.
5. References.
Alholjailan, M.I. (2012). Thematic Analysis: A critical review of its process and evaluation.
West East Journal of Social Sciences, 1(1), 39-47.
Boyatzis, R. E. (1998). Transforming qualitative information: thematic analysis and code
development. Sage.
Braun, V. & Clarke, V. (2006). Using thematic analysis in psychology. Qualitative Research in
Psychology, 3, 77-101.
Bree, R. & Gallagher, G. (2016). Using Microsoft Excel to code and thematically analyse
qualitative data: a simple, cost-effective approach. All Ireland Journal of Teaching and
Learning in Higher Education (AISHE-J), 8(2), 2811-28114.
AISHE-J
Volume 8, Number 3 (Autumn 2017)
33514
Clarke, V. & Braun, V. (2013) Teaching thematic analysis: Overcoming challenges and
developing strategies for effective learning. The Psychologist, 26(2), 120-123.
Divan, A., Ludwig, L., Matthews, K., Motley, P. & Tomlienovic-Berube, A. (2017). A survey of
research approaches utilised in The Scholarship of Learning and Teaching publications.
Teaching & Learning Inquiry,[online] 5(2), 16.
Javadi, M. & Zarea, M. (2016). Understanding Thematic Analysis and its Pitfalls. Journal Of
Client Care, 1 (1) , 33-39.
Nowell, L. S., Norris, J. M., White, D. E., & Moules, N. J. (2017). Thematic Analysis: Striving to
Meet the Trustworthiness Criteria. International Journal of Qualitative Methods, 16 (1), 1-13.
O’Cathain, A., & Thomas, K. J. (2004). “Any other comments?” Open questions on
questionnaires – a bane or a bonus to research? BMC Medical Research Methodology, 4, 25.
Rosenthal, M. (2016). Qualitative research methods: Why, when, and how to conduct
interviews and focus groups in pharmacy research. Currents in Pharmacy Teaching and
Learning, 8(4), 509-516.
Rowland, S.L. & Myatt, P.M. (2014). Getting started in the scholarship of teaching and
learning: a "how to" guide for science academics. Biochemistry & Molecular Biology
Education, 42(1), 6-14.
Vaismoradi, M., Turunen, H. & Bondas, T. (2013). Content analysis and thematic analysis:
Implications for conducting a qualitative descriptive study. Nursing and Health Sciences,
15(3), 398-405.
Download