Types of Research Design - Oncourse

advertisement
Types of Research in Language Education
Larry Mikulecky
Indiana University-Bloomington
If you are in this class, chances are you are pursuing a doctorate and will
quite soon be in a position to serve on the research committees of graduate
students and sometimes direct their research as their major advisor. In
addition, you will be expected to review the research others submit to journals
and conferences as part of the service you contribute to the profession.
Though you may be partial to a particular research topic and/or approach, you
are unlikely to have the luxury of restricting yourself to only what pleases you.
Your employer, your students, and your profession will expect you to
contribute to judging several sorts of research quality and being able to offer
suggestions to improve many kinds of research.
There is no consensus about the best way to classify the various types,
approaches, and methods used to do educational research in general and
language research in particular. For example, Mason & Bramble (1997) list
five overlapping categories of research:
1)
2)
3)
4)
5)
Descriptive research;
Experimental and quasi-experimental;
Qualitative;
Historical; and
Evaluation and policy studies.
Within each of these broad categories, they include anywhere from three to
six sub-types. They point out that studies sometimes are of mixed
methodology using more than one type of research approach to answer key
questions (pp. 37-44).
Gay and Airasian (2000) list six types of research (i.e. historical,
qualitative, descriptive, correlational, causal-comparative, and experimental).
Labels like correlational and causal-comparative, which are sub-categories for
Mason and Bramble, become full categories in their own right for Gay and
Airasian. Evaluation studies and policy studies aren’t addressed by Gay and
Airasian. As with Mason and Bramble, Gay and Airasian do list several subtypes under each of their major classifications. None of the authors above list
two important types of research (i.e. studies to develop new theory and
studies that synthesize and meta-analyze existing research).
Gay and Airasian try to clarify the relationships among the types of
research they address in the diagram below. As you examine the diagram, try
to think of studies you know of that use each type of research approach. If
you haven’t encountered (or can’t clearly remember) a study that fits within
each of Gay and Airasian’s categories, put a check mark next to the term. We
will be using these check marks (as well as check marks you make next to
other terms in this hand-out) as starting places for small group and full class
discussions. Take a moment, now, to place check marks next to or make
note of any terms on this page or the diagram on the next page for
which you are unable to think of a concrete study example.
1
2
Sutter (1998) takes another approach to describing the various sorts of
educational research when he writes:
“Labelling educational research as a “type” is not as important as
understanding the implications of a study’s most distinctive features. I
believe that describing the distinctive features of educational research
and thus avoiding contrived artificial typologies, captures the
complexity and does not place imposing restraints on researchable
questions.” (p. 86).
He goes on to suggest that there are six distinguishing features that should
be taken into account. These are:
1)
2)
3)
4)
5)
6)
Quantitative versus Qualitative,
Descriptive versus Inferential,
True Experiment versus Quasi-experimental,
Causal Comparative versus Correlational
Single-Subject versus Group, and
Teacher (sometimes call action research) versus Traditional.
Sutter’s distinguishing characteristics cover the categories and subcategories of the authors discussed earlier. They also introduce a new
research type (i.e. teacher or action research).
In this course and throughout your professional life you will be
expected to critically read and provide feedback about all the sorts of
research. As you do this, three questions remain central. These are:
1) What kind of evidence is sufficient and can be trusted for this kind
of study?
2) How far can one extend beyond the evidence in this study to
drawing conclusions and making recommendations?
3) Does this study link to previous research and push the edge of
knowledge?
Throughout this semester, you will be examining several sorts of research
and providing feedback as if you had already completed your doctorate. This
will occur in several ways.
1)
After seeing modelling about how to review manuscripts in class, you
will review manuscripts as if you were a journal reviewer or a faculty
member providing professional feedback to the research of a student
or colleague. Some of the material you will review is already
published and some is in manuscript form and hasn’t yet been
published.
2)
You will compare the strengths and weaknesses you identify in your
reviews with those identified by other students and also by actual
3
journal reviewers. In addition, you (with 1-2 other students) will be
both presenting someone else’s doctoral dissertation to the class and
sometimes functioning as a review committee asking questions about
the dissertation and making suggestions.
3)
You will design a small pilot study that may serve as a steppingstone to your dissertation. This study should link to and go beyond
existing research. Examples are:
A) developing and trying out a means for gathering evidence;
B) doing some initial analysis of data to see if full analysis is
warranted, or
C) Identifying where the weaknesses are in previous studies and
suggesting ways to overcome these weaknesses.
You will be presenting the results of your pilot study to a small
group of your colleagues in this course during the last two class
sessions.
4
Glossary of Some Key Research Terms*
Action research/Teacher research: These studies usually involve a teacher
gathering evidence to study her or his own teaching, to improve skills, or to
understand the learning process of students. Such research is mainly for the
teacher’s own use, though it can be more widely useful if it connects to
previous research and extends our conceptual knowledge. This research is
sometimes criticized for having inadequate research design and lacking
trustworthy evidence.
Case Study: These studies involve collecting data from several sources over
time, generally from only one or a small number of cases. It usually provides
rich detail about those cases, of a predominantly qualitative nature.
Causal-comparative research. These studies attempt to establish causeeffect relationships among the variables of the study. The attempt is to
establish that values of the independent variable have a significant effect on
the dependent variable. This type of research usually involves group
comparisons. For example, a causal-comparative study might attempt to
identify factors related to the drop-out problem in a particular high school
using data from records over the past ten years; to investigate similarities and
differences between such groups as smokers and nonsmokers, readers and
non readers, or delinquents and non-delinquents, using available data. In
causal-comparative research the independent variable is not under the
experimenter’s control, that is, the experimenter can't randomly assign the
subjects to a group, but has to take the values of the independent variable as
they come. The dependent variable in a study is the outcome variable.
Correlational research: These studies attempt to determine whether and to
what degree, a relationship exists between two or more quantifiable
(numerical) variables. However, it is important to remember that just because
there is a significant relationship between two variables it does not follow that
one variable causes the other. When two variables are correlated you can use
the relationship to predict the value on one variable for a subject if you know
that subject's value on the other variable. Correlation implies prediction but
not causation. The investigator frequently uses the correlation coefficient to
report the results of correlational research. Multiple regression studies
determine the ability of several different variables taken together and in
various sequences to predict a new variable (i.e. study time, quality of
instruction, home background used to predict language learning gains).
Descriptive research: These studies involve collecting data in order to test
hypotheses or answer questions regarding the subjects of the study. In
contrast with the qualitative approach the data are numerical. The data are
typically collected through a questionnaire, an interview, or through
observation. In descriptive research, the investigator reports the numerical
results for one or more variables on the subjects of the study. Examples
*
Many of these definitions draw upon course material developed by Jonathan Plucker at
Indiana University and John Wasson at Morehead State University.
5
include public opinion surveys, fact-finding surveys, job descriptions, surveys
of the literature, documentary analyses, anecdotal records, critical incident
reports, test score analyses, normative data, description of the type and age
of computers in rural schools. Policy makers often rely on this type of
research to inform and justify their decisions.
Evaluation: This might be an evaluation of a curriculum innovation or
organisational change. An evaluation can be formative (designed to inform
the process of development) or summative (to judge the effects). Often an
evaluation will have elements of both. If an evaluation relates to a situation in
which the researcher is also a participant it may be described as “teacher
research” or “action research”. Evaluations will often make use of case study
and survey methods and a summative evaluation will ideally also use
experimental methods. For an evaluation to “push the edges of knowledge” it
must go beyond “Did this work when we tried it?” to provide information that
extends our knowledge of “why” and “how” something worked.
Experimental Research: These studies investigate possible cause-andeffect relationships by exposing one or more experimental groups to one or
more treatment conditions and comparing the results to one or more control
groups not receiving the treatment (random assignment being essential).
Issues of generalisablity (often called ‘external validity’) are usually important
in an experiment, so the same attention must be given to sampling, response
rates and instrumentation as in a survey (see below). It is also important to
establish causality (‘internal validity’) by demonstrating the initial equivalence
of the groups (or attempting to make suitable allowances), presenting
evidence about how the different interventions were actually implemented and
attempting to rule out any other factors that might have influenced the result.
In education, these studies are difficult to implement because of logistical
problems and sometimes ethical problems.
Qualitative approach Qualitative approaches involve the collection of
extensive narrative data in order to gain insights into phenomena of interest.
Data analysis includes the coding of the data and production of a verbal
synthesis (inductive process). Trustworthiness is usually achieved by
supporting conclusions with information triangulated from several different
sources and/or the same source over time. Inter-rater agreement about what
goes into qualitative categories is usually called for.
Quantitative approach: Quantitative approaches involve the collection of
numerical data in order to explain, predict, and/or control phenomena of
interest, data analysis is mainly statistical (deductive process).
Trustworthiness is usually achieved by establishing several sorts of validity
and reliability for the measures being used to create the numerical data.
Quasi-experimental research: These studies are not able to make use of
random assignment, though control groups are usually expected. Attempts
are made to demonstrate that the treatment and control groups are somewhat
equivalent and to statistically account for beginning differences if the groups
aren’t equivalent. Less credible quasi-experiments do not include control
groups and simply document change for participants receiving a particular
6
treatment. Like true experiments, there is a expectation that acceptable
validity and reliability have been established for the information collected.
Research synthesis and meta-analysis: A research synthesis is an attempt
to summarise or comment on what is already known about a particular topic.
The researcher brings different sources together, synthesising and analysing
critically, thereby creating new knowledge (i.e. the studies together tell us
more than individual studies alone). High quality studies go to great lengths to
ensure that all relevant sources (whether published or not) have been
included. Details of the search strategies used and the criteria for inclusion
must be made clear. A systematic review will often make a quantitative
synthesis of the results of all the studies, for example by meta-analysis (i.e.
using statistical techniques to compile data from similar studies).
Survey: An empirical study that involves collecting information from a larger
number of cases, perhaps using questionnaires. A survey might also make
use of already available data, collected for another purpose. A survey may be
cross-sectional (data collected at one time) or longitudinal (collected over a
period). Because of the larger number of cases, a survey will generally
involve some quantitative analysis. Issues of generalisablity are usually
important in presenting survey results, so it is vital to report how samples were
chosen, what response rates were achieved and to comment on the validity
and reliability of any instruments used.
7
References
Gay, L. & Airasian, P. (2000). Educational Research: Competencies for
Analysis and Application, Sixth Edition, Merrill/Prentice Hall
Mason, E. & Bramble. W. (1997). Research in Education and the Behavioral
Sciences. Madison, WI, Brown and Benchmark.
Sutter, W. N. (1998). Primer of Educational Research. Boston, MA. Allyn
and Bacon.
8
Download