Research Problem Statement Construct, Concept and Variables Research Process The research process: • Characteristics of research: – Proprietary vs. scholarly research – Research is based on curiosity and asking questions (creativity) – Research is a systematic process – Research is potentially replicable – Conceptualization, planning and designing research, methodologies for conducting research, analyzing and interpreting data, reconceptualization – Research is reflexive and self critical – Research is cumulative and self-corrective – Research is cyclical • Research paradigms: – Physical sciences – Humanities – Social sciences Assumption Ontological assumption Epistemological assumption Axiological assumption Methodological assumption Rhetorical assumption Question What is the nature of reality What is the relationship of the researcher to that being researched Positivist paradigm Singular Objective Independent Naturalistic paradigm Multiple Intersubjective Interdependent What is the role of values in the research process What is the process of research Value-free Unbiased Value-laden Biased Deduction Search for cause and effect relationships between variables Static design Researcher-controlled setting Quantitative methods Context-free generalizations Goals of explanation, prediction and control Induction Holistic understanding of patterns of behavior Emergent behavior Natural setting Qualitative methods Context-bound findings Goals of understanding and social change What is the language of Formal research reports? Impersonal voice Informal Personal voice Concept • An abstraction encompassing observed events; a word that represents the similarities or common aspects of objects or events that are otherwise quite different from one another. • The purpose of a concept is to simplify thinking by including a number of events (or the common aspects of otherwise diverse things) under one general heading (Ary 1985). • Chair, dog, tree, liquid, a doughnut, etc… Construct • Constructs are the “highest highest-level abstractions” of complicated objects and events, created by combining concepts and less complex constructs. – used to account for observed regularities and relationships, and to summarize observations and explanations (Ary 1985). • A concept with added meaning of having been deliberately and con consciously invented or seriously adopted for a special scientific purpose. 1) it enters into theoretical schemes and is theoretical related in various ways to other constructs. 2) it is defined and specified so that it may be observed or measured (Kerlinger 1986). Construct • Scientists measure things in three classes: direct observables, indirect observables (not experienced or observed first hand), and constructs. • These constructs are defined as constructs theoretical creations based on observations but cannot be observed directly or observed indirectly (Kaplan 1964). • Motivation, visual acuity, justice, problem solving ability, …not a doughnut, …but hunger. problem- Operational Definition • It describes meaning to a concept or construct by specifying the operations that must be performed in order to measure or manipulate the concept, as the data collected during research is in terms of observable events (Ary 1985). • It defines or gives meaning to a variable by spelling out what the investigator must do to measure it (Kerlinger 1986). • “Operational definitions are essential to research because they permit investigators to measure abstract concepts and constructs and permit scientists to permit move from the level of constructs and theory to the level of observation” (Ary 1985). Operational Definition • Two Types of Operational Definitions: • Measured Operational Definition: Operations by which investigators may measure a concept. • Experimental Operational Definition: Steps taken by a researcher to produce certain experimental conditions. Operational Definition • Examples of an Operational Definition: • Measured Operational Definition: An actual (score) value from a test or questionnaire the researchers would develop to measure “hunger.” • Experimental Operational Definition: A manipulated scenario to produce the condition of “hunger.” (such as preventing the subject from consuming anything for x number of hours) x- Variable • Characteristics or attributes of an object, individual or organization that can be measured or observed, and that varies among those objects or individuals being studied (Creswell 2002). • They possess values and levels (the dimensions on which they vary) (Sommery 1997). • “The concepts that are of interest in a study become the variable variables for s investigation (Ary 1985).” Different Kinds of Variables • Dichotomous: Two valued variables. Example: Sex (male/female) Two• Polytomous: Multiple values for variables. Example: Religion (Catholicism, Islam, Judaism, Hinduism, Buddhism, etc…) • Continuous: A variable that takes on an infinite number of values within a range. Example: Height & Weight More Kinds of Variables • Independent: The variable manipulated by the experimenter (also: Experimental, Predictor, Manipulated, Antecedent, Treatment). • Active: Any variable that is manipulated by the researcher • Attribute: Any variable that cannot be manipulated by the researcher. For example, all human characteristics are attribute variables: intelligence, sex, socioeconomic status etc. • Dependent: The dependent variable is the phenomenon that is the object of study and investigation (also: Outcome, Response, Criterion, Effect). - - More Kinds of Variables • Categorical: Referred to as nominal measurements. One creates ‘categories,’ and classifies all variables that fall under this definition without rank order. All variables under the same category are considered of equal value, and not differentiated. • Latent: An unobserved ‘entity’ that “stands between” the independent variable and the dependent variable, and mediates the effect of the independent variable on the dependent variable. It is dependent on the independent variable as well as other constructs, yet still plays a role in determining the outcome (possibly: Intervening, Mediating, Hypothetical construct). More Kinds of Variables • Control: An independent variable that is measured in a study because the y potentially influence the dependent variable. It is a more clearly defined independent variable in attempts to eliminate all bias in regard s to its effects on the dependent variable. (Keeps the study in check). they clearly regards • Confounding: Variables not actually measured or observed in a study, yet they exist, and its influence cannot be directly detected or understood in a study. One becomes aware of a confounding variable at the end of a study, they realize that there is an effect that was not measured or accounted for, but should be addressed. Concepts, Construct and Variables • Concepts (Oxford definition) general notation, abstract ideas. • Construct (Oxford definition) make by fitting together; build; thing constructed; especially by the mind. • Variables (Oxford definition) a situation, number or quantity that can vary or be varied • Constant (Oxford definition) a number or quantity that does not vary Concepts, Construct • Concepts are abstract ideas which have been "defined" according to particular characteristics or generalizations (constructs) about them. • A construct is based on concepts, or can be thought of as a conceptual model that has measurable aspects. • This will allow the researcher to "measure" the concept and have a common acceptable platform when other researches do a similar research. Concepts, Construct • E.g • Measuring advertising effectiveness is an construct, and concepts related would be brand awareness and consumer behavior. • Pain is a concept, a theoretical model of pain would be a construct, and a pain assessment tool would give a measurable variable. Concepts, Construct • An idea is a plan, suggestion, or possible course of action. ex...I really like the idea of helping people. A concept is an idea or abstract principle. ex....She added that the concept of arranged marriages is misunderstood in the West. • Science uses: i. Concepts ii. Links concepts by propositions iii. Testing theory with observable evidence iv. Publication of definitions and procedures v. Control of alternative explanations vi. Unbiased selection of evidence vii. Reconciliation between theory and observation Concepts, Construct • A concept is a verbal abstraction drawn from observation of a number of specific cases • A theoretical definition explains what is meant by the concept. • Operational definitions translate the verbal concepts into corresponding variables which • can be measured. Operational definitions can be either: measured, or experimental. • Also, a variable can be either measured (e.g., surveys) or manipulated (e.g., experiment). Concepts, Construct • A construct serves the same function as a concept, but it is more abstract. • It is not characterized by a direct link between the abstraction and its observed manifestations. • For instance, “source credibility” is a construct which has been used in studying persuasion. • This term can be used in the same way as a concept, but we should recognize that we cannot directly observe different levels of source credibility in individuals. • However, we can observe the various parts which make up the construct individually, and then combine them to get some overall summary. Concepts, Construct • Constructs are built from the logical combination of a number of more observable concepts. In the case of source credibility, we could define the construct as the combination of the concepts of expertise, objectivity, and status. • Each of these concepts can be more directly observed in an individual. • We might also consider some of these terms to be constructs themselves, and break them down into combinations of still more concrete concepts What we see if we do this is a set of constructs at decreasing levels of abstraction. Only at the bottom of this hierarchy are directly observable concepts. Concepts, Construct • A scientific concept really consists of three parts: a label, a theoretical definition, and an operational definition. • The theoretical definition specifies the verbal meaning which is attached to the concept label. • We call self-defining concepts like “age” primitive terms. • Primitive terms are adequately defined by their attached concept labels. • These are the labels which appear at the bottom of the level of abstraction hierarchy. • RECOMMENDATION: explicitly specify the meaning associated with each concept, regardless of the extent to which we think the meaning is shared. Concepts, Construct • An operational definition translates the verbal meaning provided by the theoretical definition into a prescription for measurement. • Although they may be expressed verbally, operational definitions are fundamentally statements that describe measurement and mathematical operations. • An operational definition describes the unit of measurement. Examples of units of measurement are minutes (to measure time), word counts (to measure newspaper coverage of a particular event), percent correct responses, etc. • An operational definition adds three things to the theoretical definition. • Operationalization is to take a fuzzy concept, such as 'helping behavior', and try to measure it by specific observations, e.g. how likely are people to help a stranger with problems. Concepts, Construct • An operational definition specifies the level of measurement. • Levels of measurement can range from the simple nominal variables which only make distinctions between categories like “present or absent” or “yes or no”; to ordinal variables which contain some information about the quantity (“more or less”) of the concept present, but have no real measurement scales; to continuous variables which have real scale points which are equally spaced, and which can take on any value. Concepts, Construct • The operational definition must be very closely associated with the theoretical definition. • It must state clearly how observations will be made so they will reflect as fully as possible the meaning associated with the verbal concept or construct. • The operational definition must tell us how to observe and quantify the concept in the “real world”. • This connection between theoretical and operational definitions is quite critical. • This connection establishes the validity of the measurement. • The amount of validity in measurement is proportional to the extent to which we actually measure what we intend to measure, that is, the degree to which the operational definition and the theoretical definition correspond. Relationships between variables/concepts • Null relationships • Covariance relationships: – In a covariance relationship, changes in the values of one variable (the measured concept) are associated with changes in the values of the other variable. – That is, the variables shift values simultaneously, or covary. – This does not mean that one concept is the cause and the other is the effect. – A cause-effect relationship between concepts requires more than just covariance, as we’ll see shortly. – A covariance relationship is typically diagramed with a curved, double-headed arrow between the concept Relationships between variables/concepts – Two variables which are related proportionally can covary either positively or negatively. – While covariance relationships can provide prediction, they can’t provide explanation of the relationship. – Spurious relationships (artifacts): two variables may covary because they are both the effects of a common cause. The unobserved, but real, causal variable sometimes is called a confounding variable, since it may mislead us by producing the appearance of a relationship between the observed variables. – Control variables Relationships between variables/concepts • Causal relationships: – Causality means that a change which occurs in one variable (the cause) brings about a change in another variable (the effect). – Alternative terms for cause and effect variable are independent variable and dependent variable. – This terminology is based on the logic that in a causal relationship the state of one variable (the effect) depends on the state of the other (the cause). – The state of the cause is independent of the state of the effect variable. Relationships between variables/concepts • There is a critical difference between covariance and causality: • Covariance means that a change in one variable is associated with a change in the other variable; causality requires that a change in one variable creates the change in the other. • In other words, covariance alone does not imply causality. Relationships between variables/concepts – Covariance is only one of four conditions which must be met before we can state that a relationship is causal: • • • • Spatial contiguity Covariance (necessary but not sufficient condition) Temporal ordering Necessary connection (This necessary connection is a statement which specifies why the cause can bring about a change in the effect. It is the logical statement of the process or mechanism by which the two variables are related to one another in a cause-effect relationship. SOME DEFINITIONS OF CONSTRUCTS PERMISSIVENESS: • Oxford def. a) tolerant; liberal; b) giving permission • Experimental def. Extending the boundaries of acceptable findings. • Measured def. Confining the boundaries of acceptable findings. REINFORCEMENT: • Oxford def. Strengthen or support, especially with additional personnel, material etc. • Experimental def. To build credibility by strengthening your research findings. • Measured def. To build structural credibility to strengthen your research findings. READING ABILITY: • Oxford def. none • Experimental def. To understand through written text the research findings. • Measured def. same as above SOME DEFINITIONS OF CONSTRUCTS ACHIEVEMENT: • Oxford def. a) something achieved b) act of achieving achieve: a) attain by effort acquire; gain earn b) accomplish • Experimental def. To accomplish what you have set out to prove. • Measured def. To put your research findings into a written format to be used as documentation. INTERESTS: • Oxford def. a) curiosity, concern b) quality existing curiosity c) note worthiness, importance 2) subject, hobby in which one is concerned 3) advantage or profit 4) self interest, excite the curiosity or attention to take a personal interest. • Experimental def. An educational topic that concerns you and is worthy of your research. • Measured def. same as above NEEDS: • Oxford def. archaic of necessity, requirement • Experimental def. The requirements of research in order for it to be valid. • Measured def. The requirements for charting the research findings in order to give documented support. SOME DEFINITIONS OF CONSTRUCTS TRANSFER OF TRAINING: • Oxford def. none • Experimental def. To give another researcher the information needed so that they can continue researching from where you left off. • Measured def. same as above LEADERSHIP: • Oxford def. A person that leads or is followed by others. • Experimental def. A person who has the ability to direct others through a research experiment, and who will set the tone of the research. • Measured def. A person who will design the way in which the research findings will be documented. CLASS ATMOSPHERE: • Oxford def. none • Experimental def. The tone of a class setting that will allow for similar testing conditions. • Measured def. same as above SOME DEFINITIONS OF CONSTRUCTS DELINQUENCY: • Oxford def. Failing in one’s duty • Experimental def. Failing to plan out the way you are going to conduct you research so that your findings are valid. • Measured def. Failing to document your findings in a way that another researcher can duplicate your research findings. ORGANIZATIONAL CONFLICT: • Oxford def. none • Experimental def. Lack of agreement within a research group. • Measured def. Lack of consistent findings and charting. SELF-OTHER-ATTITUDE: • Oxford def. none • Experimental def. Ability to work together for the good of the research • Measured def. same as above SOME DEFINITIONS OF CONSTRUCTS CONFORMITY: • Oxford def. Accordance with established practice, agreement • Experimental def. To follow a specific plan that has already been established and agreement as a research team to stick to a research method that was agreed upon. • Measured def. To chart the research findings in a way that would best support your findings and based on a method that was used prior by another research team when studying the same subject area. Variables • A variable is something that changes. • It changes according to different factors. • Some variables changes easily, like the stockexchange value, while other variables are almost constant, like the name of someone. • Researchers are often seeking to measure variables. • The variable can be a number, a name or anything where the value can change. Variables • An example of a variable is temperature. • The temperature varies according to other variable and factors. • You can measure different temperature inside and outside. • If it is a sunny day, chances are that the temperature will be higher than if it's cloudy. • Another thing that can make the temperature change is whether something has been done to manipulate the temperature, like lighting a fire in the chimney. Variables • A variable is any entity that can take on different values. • Anything that can vary can be considered a variable. • For instance, age can be considered a variable because age can take different values for different people or for the same person at different times. Variables • Variables are not always 'quantitative' or numerical. • The variable 'gender' consists of two text values: 'male' and 'female'. • We can, if it is useful, assign quantitative values instead of (or in place of) the text values, but we don't have to assign numbers in order for something to be a variable. • It's also important to realize that variables aren't only things that we measure in the traditional sense. • For instance, in much social research and in program evaluation, we consider the treatment or program to be made up of one or more variables (i.e., the 'cause' can be considered a variable). • An educational program can have varying amounts of 'time on task', 'classroom settings', 'student-teacher ratios', and so on. Variables • Variables may have the following characteristics: – Period: When it starts and stops. – Pattern: Daily, weekly, ad-hoc, etc. – Detail: Overview through to 'in depth'. – Latency: Time between measuring dependent and independent variable (some things take time to take effect). Variables • In research, you typically define variables according to what you're measuring. • The independent variable is the variable which the researcher would like to measure (the cause), while the dependent variable is the effect (or assumed effect), dependent on the independent variable. • These variables are often stated in experimental research, in a hypothesis, e.g. "what is the effect of personality on helping behavior?“ Variables • In explorative research methodology, e.g. in some qualitative research, the independent and the dependent variables might not be identified beforehand. • They might not be stated because the researcher does not have a clear idea yet on what is really going on. • The independent variable, also known as the manipulated variable, lies at the heart of any quantitative experimental design. • A researcher manipulates an independent variable, to influence a dependent variable, or variables. • There may be more than two dependent variables in any experiment. Variables • The independent variable is what you (or nature) manipulates -- a treatment or program or cause. • The dependent variable is what is affected by the independent variable -- your effects or outcomes. • For example, if you are studying the effects of a new educational program on student achievement, the program is the independent variable and your measures of achievement are the dependent ones. Variables • The independent variable (IV) is often thought of as our input variable. • It is independent of everything that occurs during the experiment because once it is chosen it does not change. • In our experiment on college performance, we chose two groups at the onset, namely, those with work experience and those without. • This variable makes up our two independent groups and is therefore called the independent variable. Variables • The dependent variable (DV), or outcome variable, is dependent on our independent variable or what we start with. • In this study, college grades would be our dependent variable because it is dependent on work experience. • If we chose to also look at men versus women, or older students versus younger students, then these variables would be other independent variables and the outcome, our dependent variable (college grades), would be dependent on them as well. • Remember that whatever is the same between the two groups is considered a constant because they do not vary between groups but rather remain the same and therefore do not affect the outcome of each group differently Variables • Descriptive variables are those that which will be reported on, without relating them to anything in particular. • Categorical variables result from a selection from categories, such as 'agree' and 'disagree'. Nominal and ordinal variables are categorical. • Numeric variables give a number, such as age. • Discrete variables are numeric variables that come from a limited set of numbers. They may result from , answering questions such as 'how many', 'how often', etc. • Continuous variables are numeric variables that can take any value, such as weight. • Extraneous variables are additional variables which could provide alternative explanations or cast doubt on conclusions. Variables • Researchers must be aware that variables outside of the independent variable(s) may confound or alter the results of a study. • Confounding variables are variables with a significant effect on the dependent variable that the researcher failed to control or eliminate sometimes because the researcher is not aware of the effect of the confounding variable. • The key is to identify possible confounding variables and somehow try to eliminate or control them. Variables • A confounding variable, also known as a third variable or a mediator variable, can adversely affect the relationship between the independent variable and dependent variable. • This may cause the researcher to analyze the results incorrectly. • The results may show a false correlation between the dependent and independent variables, leading to an incorrect rejection of the null hypothesis. Variables • If, for instance, we had two groups in the above mentioned study but did not control for age then age itself may be a confound. Imagine comparing students with work experience with a mean age of 40 with students without work experience and a mean age of 18. • Could we reasonably say that work experience caused the student to receive higher grades? • This extraneous variable can play havoc on our results as can any intervening variable such as motivation or attention. • Addressing confounds before they alter the results of your study is always a wise decision Variables Every Problem Needs Further Delineation • To comprehend fully the meaning of the problem, the researcher should eliminate any possibility of misunderstanding by – Stating the hypotheses and/or research questions: Describing the specific hypotheses being tested or questions being asked. – Delimiting the research: Fully disclosing what the researcher intends to do and, conversely, does not intend to do. – Defining the terms: Giving the meanings of all terms in the statements of the problem and subproblems that have any possibility of being misunderstood. – Stating the assumptions: Presenting a clear statement of all assumptions on which the research will rest. • These matters facilitate understanding of the research – called the setting of the problem Stating the Hypotheses and/or Research Questions • Hypotheses are tentative, intelligent guesses posited for the purpose of directing one’s thinking toward the solution of the problem • Necessary in searching for relevant data and in establishing a tentative goal • Hypotheses are neither proved nor disproved. They are nothing more than tentative propositions set forth to assist in guiding the investigation of a problem or to provide possible explanations for the observations made Accept/Reject Hypotheses • Hypotheses have nothing to do with proof • Their acceptance or rejection is dependent on what the data – and the data alone – ultimately reveal • Hypotheses may originate in the subproblem, could be 1 to 1 • Hypothesis provides a position from which a researcher begins to initiate an exploration of problem and subproblems and checkpoints to test the findings that the data reveal Accept/Reject Hypotheses • If the data do not support the research hypothesis, don’t be disturbed – it merely means that the educated guess about the outcome of the investigation was incorrect • Frequently, rejected hypotheses are a source of genuine and gratifying surprise – truly made unexpected discovery • Another type of hypothesis is the null hypothesis Null Hypothesis • It is an indicator only • Reveals some influences, forces, or factors that have resulted in a statistical difference or no such difference • Most researches stop at this point – getting off at mezzanine instead down to the basement where the foundations are Null Hypothesis Dynamics • If null hypothesis shows the presence of dynamics, then the next logical questions are as follows: – What are these dynamics? – What is their nature? – How can they be isolated and studied? • For example, let’s say that a team of social workers believe that one type of after-school programme for teenagers (we’ll call it Programme A) is more effective than another programme (we’ll call it Programme B) in terms of reducing high school dropout rates. Null Hypothesis Dynamics • The null hypothesis stating that there will be no difference in the high school graduation rates of teenagers enrolled in Programme A and those enrolled in Programme B has been rejected – encouraging news – it is mezzanine conclusion • What specifically were the factors within the programme that cause the null hypothesis to be rejected? • These are fundamental questions – will uncover facts that may lie very close to the discovery of new substantive knowledge – the purpose of all research Delimiting the Research • Know PRECISELY what the researcher intends to DO and does NOT intend to do • What the researcher intends to do is stated in the problem statement • What the researcher is not going to do is in the delimitations • The researcher can easily be beguiled (deceived, cheated) by discovering interesting information that lies beyond the precincts of the problem under investigation • Only a researcher who thinks carefully about the problem and its focal centre can distinguish between what is relevant and what is not relevant to the problem • All irrelevancies to the problem must be firmly ruled out in the statement of delimitations Defining The Terms • Without knowing explicitly what a term means, we cannot evaluate the research or determine whether the researcher has carried out what was proposed in the problem statement • Need not necessarily agree with such a definition, but as long as we know what the researcher means when using the term, we are able to understand and appraise it appropriately • A formal definition contains three parts: (a) the term to be defined; (b) the genera, or the general class to which the concept being defined belongs; and (c) the differentia, the specific characteristics or traits that distinguish the concept being defined from all other members of the general classification Defining The Terms • To make the software more USER-FRIENDLY? • What is the relationship between the user interface metric and user acceptance? • The researcher must be careful to avoid circular definitions, in which the terms to be defined are used in the definitions themselves • A classic example is Gertrude Stein’s “A rose, is a rose, is a rose” Stating the Assumptions • Assumptions are so basic that, without them, the research problem itself could not exist • Example, to determine by pretest-posttest whether one method of instruction has produced the results hypothesised • The assumptions are: – The test measures what it is presumed to measure – The teacher(s) in the study can teach effectively – The students are capable of learning the subject matter • Without these assumptions, we have no problem, no research Stating the Assumptions • Assumptions are what researchers take for granted with respect to the problem • But taking everything for granted may cause misunderstanding • If others know the assumptions a researcher makes, they are better prepared to evaluate the conclusions that result from such assumptions • Many students thought that assumption is stating the obvious • In research, try to leave nothing to chance in the hope of preventing any misunderstanding Stating the Assumptions • All assumptions that have a material bearing on the problem should be openly and unreservedly set forth. • Asking question “What am I taking for granted with respect to the problem?” will bring assumptions into clear view Importance of the Study • In dissertations or research reports, researchers frequently set forth their reasons for undertaking the study • In a research proposal, such a discussion may be especially important • Some studies seem to go far beyond any relationship to the practical world • Of such research efforts, one might asks “Of what use is it? What practical value does the study have?” • For example, the time, money, effort spent on early space exploration flights