How Do We Know What We Know about Organizational Behavior? By Uma Sekaran This book has examined the skills and knowledge that managers need to be successful in their jobs. But how do you know how much faith you should have in all the information you acquire from textbooks and management journals? Are some theories and statements more applicable than others? Even when applicable, will they apply at all times and under all circumstances? You can find answers to these important questions once you know the foundation on which theories and assertions rest. This appendix provides that foundation. It first examines why managers need to know about research, and then discusses the basis for knowledge in this field. It then looks at the research process and research design and ends with a discussion of how research knowledge affects you. WHY MANAGERS SHOULD KNOW ABOUT RESEARCH Why is it necessary for you to know about research? First, this knowledge helps you determine how much of what is offered in textbooks is of practical use to you as a manager. Second, a basic understanding of how good empirical research is done can make you an effective manager by helping you to make intelligent decisions about research proposals and reports that reach your desk. Third, it enables you to become an informed and discriminating consumer of research articles published in the management journals that you need to read to keep up with new ideas and technology. For your convenience, a list of the current academic and practitioner-oriented journals that frequently publish articles on organizational behavior is provided in Table B.1. T a b l e B . 1 A r t i c l e s J o u r n a l s w i t h O r g a n i z a t i o n a l Academic Journals Practitioner-Oriented Journals Academy of Management Journal Academy of Management Review Administrative Science Quarterly Advances in International Academy of Management Executive Business Horizons California Management Review Columbia Journal of World Business Comparative Management American Journal of Small Business Behavioral Science Group and Organization Studies Human Relations Human Resource Management Human Resource Management Review Harvard Business Review Human Resource Development Quarterly Industrial Relations Industry Week Organizational Dynamics Personnel Journal SAM Advanced Management Journal HUMAN RESOURCE PLANNING Industrial and Labor Relations Review International Journal of SLOAN MANAGEMENT REVIEW Supervision Training B e h a v i o r Management Training and Development Journal Journal of Applied Behavioral Science Journal of Applied Business Research Journal of Applied Psychology Journal of Business Journal of Business Ethics Journal of Business Research Journal of Human Resource Management Journal of International Business Studies Journal of Management Journal of Management Studies Journal of Occupational Psychology Journal of Organizational Behavior Journal of Organizational Behavior Management Journal of Vocational Behavior Organizational Behavior and Human Decision Processes Personnel Administrator Sex Roles Women in Business Understanding scientific research methods enables you to differentiate between good and appropriate research, which you can apply in your setting, and flawed or inappropriate research, which you cannot use. Moreover, knowledge of techniques such as sampling design enables you to decide whether the results of a study using a particular type of sample in certain types of organizations is applicable to your setting. Managers need to understand, predict, and control the research-oriented problems in their environment. Some of these problems may be relatively simple and can be solved through simple data gathering and analysis. Others may be relatively complex, needing the assistance of researchers or consultants. In either case, without some basic knowledge of scientific research, managers will be unable to solve the problems themselves or to work effectively with consultants. Managers need to discuss their problems with consultants in a useful way. This includes informing the problem solvers right at the start of the consulting process of any constraints (such as company records that are off-limits to outsiders) or of types of recommendations that will not be considered (such as laying off or hiring more people). Such discussions not only save time but also help the managers and researchers start off on the right foot. Managers who don’t understand the important aspects of research will not be equipped to anticipate and forestall the inevitable hurdles in manager–researcher interactions. Also, paying a consultant handsomely for a research report will not help the company unless the manager is capable of determining how much scientific value can be placed on the findings. For these and other reasons, a working knowledge of the scientific research process and research design is necessary. OUR BASIS FOR KNOWLEDGE Observation and scientific data gathering have led to some of our knowledge about management. For instance, very early on, Frederick Winslow Taylor observed, studied, experimented, and demonstrated that coal-mining operations could be more efficiently managed by changing the way men shoveled coal—changing how the shovel was handled, how the body movements were made, and so on. The era of scientific management that Taylor’s work ushered in provided much knowledge about how management could improve efficiency. This type of knowledge is not easy to come by, however, when we are examining employees’ feelings, attitudes, and behaviors. Our knowledge of organizational behavior stems instead from armchair theories, case studies, and scientific research. Armchair Theories In trying to understand organizational behavior, management experts and scholars initially resorted to armchair theorizing—theorizing based on the observation of various phenomena and behaviors in the workplace. For instance, Douglas McGregor, through observation and experience, theorized that managers have two different world views of employees. Some managers (Theory X) assume that employees are by nature lazy and not very bright, that they dislike responsibility and prefer to be led rather than to lead, and that they resist change. Other managers (Theory Y) assume that employees have the opposite characteristics. McGregor’s concept of Theory X and Theory Y managers has become a classic armchair theory. Few people either totally accept or totally dispute this theory because of the lack of hard data to either substantiate or negate this interesting notion. Armchair theories are based on natural observation with no systematic experimentation and hence are not very useful for application in organizations. Case Studies Case studies—studies that examine the environment and background in which events occur in specific organizations in a particular period of time—help us to understand behavior in those organizations at that time. For example, we could study a particular organization in depth to determine the contributing factors that led to its fast recovery after a prolonged recession. We might find several factors, including price reductions, the offering of good incentives to a highly motivated work force, and the taking of big risks. However, the findings from this one-time study of an organization offer only limited knowledge about fast recovery from recessions because the findings may not hold true for other organizations or for even the same organization at another time. The replication of case studies is almost impossible, since environmental and background factors are rarely the same from organization to organization. Most of the companies whose problems you have been asked to solve are from real cases written by management scholars who studied the companies. The solutions they found may not work for other organizations experiencing similar problems because of differences in size, technology, environment, labor force, clientele, and other internal and external factors. However, through case studies, we do gather information and gain insights and knowledge that might help us to develop theories and test them later. Scientific Research Empirical or data-based scientific research identifies a problem and solves it after a systematic gathering and analysis of the relevant data. This type of research offers in-depth understanding, confidence in the findings, and the capability of applying the knowledge gained to similar organizations. Scientific research is the main focus of this appendix. SCIENTIFIC INQUIRY Scientific inquiry involves a well-planned and well-organized systematic effort to identify and solve a problem. It encompasses a series of well-thought-out and carefully executed activities that help to solve the problem—as opposed to the symptoms—that is identified. Purposes of Scientific Research: Applied and Basic Research Scientific inquiry can be undertaken for two different purposes: to solve an existing problem that a particular organization faces, or to examine problems that organizations generally encounter and to generate solutions, thereby expanding the knowledge base. Research undertaken to solve an existing problem in a specific setting is applied research. In this type of research, the findings are immediately applied to solve the problem. Many professors acting as consultants to organizations do applied research. Research undertaken to add information to our existing base of knowledge is basic research. A large number of issues are of common interest to many organizations—for example, how to increase the productivity of a diverse workforce or how to eradicate sexual harassment in the workplace. The knowledge gained from research on such general issues can become useful later for application in organizational settings, but that is not the primary goal of basic research. The goal is to generate knowledge with which to build better theories that can be tested later. Basic research is often published in academic journals. The Two Faces of Science: Theory and Empirical Research Theory and empirical research are the two faces of science. Organizations benefit when good theories are developed and then substantiated through scientific research, because the results can then be confidently used for problem solving. THEORY A theory is a postulated network of associations among various factors that a researcher is interested in investigating. For example, given what has been published thus far, you might theorize that self-confident employees perceive their work environment positively, which fosters their productivity, which in turn generates more profits for the company. In constructing this theory, you have postulated a positive relationship between (1) the self-confidence of employees and their positive attitude toward their work environment, (2) their attitude toward the work environment and their productivity, and (3) their productivity and the company’s profits. No doubt, this theory appeals to common sense; but in order to establish whether or not it holds true, we need to actually test it in organizations. Thus, theories offer the basis for doing scientific, data-based research; and the theories and research together add to our knowledge. Conducting empirical research without the basis of sound theories does not steer us in the right direction, and building theories without empirically testing them limits their value. The usefulness of good theories cannot be overstated. A good theory is formulated only after a careful examination of all the previous research and writings on the topic of interest, so that no factor already established as important is inadvertently omitted. Theory building offers unique opportunities to look at phenomena from different perspectives or to add new dimensions to existing ways of examining a phenomenon. New insights and creative ideas for theory building can come through personal observation, through intuition, or even through informal discussions with employees. Testable theories are theories whose hypothesized relationships among measurable variables can be empirically tested and verified. When tested and substantiated repeatedly, such theories become the foundation on which subsequent theory building progresses. The next issue of interest is how theories are affirmed through empirical research. EMPIRICAL RESEARCH As we have just seen, theories are of no practical use unless we have confidence that they work and can be applied to problem solving in organizational settings. Empirical research allows us to test the value of theories. Empirical research is research that involves identifying the factors to be studied, gathering the relevant data, analyzing them, and drawing conclusions from the results of data analysis. It could involve simple qualitative analysis of the data, or it could be more complex, using a hypothetico-deductive approach. In qualitative analysis, responses to open-ended questions are obtained and meaningfully classified, and certain conclusions are drawn. In the hypothetico-deductive approach, a problem is identified, defined, and studied in depth; then, a theory is formulated; from that theory, testable hypotheses are generated; next, a research design is developed, relevant data are gathered and analyzed, results are interpreted, and conclusions (or deductions) are drawn from the results. Figure B.1 illustrates this approach. To be called “scientific,” research should conform to certain basic principles. It should be conducted objectively (without subjective biases). It should have a good and rigorous design (which we will examine shortly). It should be testable; that is, the conjectured relationships among factors in a setting should be capable of being tested. It should be replicable; that is, the results must be similar each time similar research is conducted. Finally, the findings should be generalizable (applicable to similar settings). It goes without saying, then, that scientific research offers precision (a good degree of exactitude) and a high degree of confidence in the results of the research (i.e. the researcher can say that 95 percent of the times, the results generated by the research will hold true, with only a 5 percent chance of its not being so). THE RESEARCH PROCESS The research process starts with a definition of the problem. To help define the problem, the researcher may interview people and study published materials in the area of interest in order to better understand what is happening in the environment. After defining the problem in clear and precise terms, the researcher develops a theoretical framework, generates hypotheses, creates the research design, collects data, analyzes data, interprets results, and draws conclusions. Problem Definition The first job for the researcher is to define the problem. Often, however, it is difficult to precisely state the specific research question to be investigated. The researcher might simply know the broad area of interest—for instance, discrimination—without being clear about which aspect of discrimination to study. In order to focus on the issue to be investigated, the researcher might need to collect some preliminary information that will help to narrow down the issue. Such information can be obtained by interviewing people in organizations and by doing a literature survey. For example, employees of different gender, race, age, physical ability, and the like may be interviewed to determine the specific aspect of discrimination on which to focus. These interviews also provide insight into what the employees (rather than the researchers) consider important. The literature survey ensures that no pertinent variable is inadvertently omitted and that there is a credible and defensible basis for the research to be done. The researcher conducts an exhaustive search for all the published work in the area of interest to determine what research has been done thus far in the particular area and with what results. The search consumes a lot of time, as one must wade through several psychological, sociological, anthropological, and other relevant journals. With all this information in hand, the researcher is now ready to define the problem. A well-defined, precise problem statement is a must for any study. The problem definition for the broad topic of discrimination could be this: What are the important factors that contribute to employees’ beliefs that they are being discriminated against by their immediate supervisor in cross-gender or cross-racial supervisor–employee relationships? Theoretical Framework The next step is to develop a theoretical framework for the study. It involves focusing on the pertinent variables for the study and discussing the anticipated or theorized network of connections among the variables. For the discrimination problem, the framework might identify three factors related to employees’ beliefs that they were discriminated against by the supervisor: (1) the level of mutual trust that is perceived by the employee to exist between the supervisor and employee (high to low), (2) the manner in which the supervisor offers performance feedback to the employee (in a forthright and helpful manner rather than in a derogatory and hurtful way), and (3) the extent to which the supervisor plays the role of mentor to the employee (training the subordinate and promoting the person’s interests in career advancement to being indifferent toward the employee’s career progress). A network of logical connections among these four variables of interest to the study—discrimination (the dependent variable) and trust, performance feedback, and mentoring (the three independent variables)—can then be formulated. These connections with the anticipated nature and direction of the relationships among the variables are postulated in the theoretical framework. Hypotheses On the basis of the theoretical framework, the researcher next generates hypotheses. A hypothesis is a testable statement of the conjectured relationship between two or more variables. It is derived from the connections postulated in the theoretical framework. An example of a hypothesis is this: The more the employee perceives the supervisor as performing the mentoring role, the less the employee will feel discriminated against by the supervisor. The statement can be tested through data gathering and correlational analysis to see if it is supported. Research Design The next step in the research process is research design. Because this step is complex, it is covered in a separate section of the chapter, after the research process. Data Collection After creating the research design, the researcher must gather the relevant data. In our example of the discrimination problem, we would collect data on the four variables of interest from employees in one or more organizations, we would obtain information about their race and gender and that of their supervisors, and we would seek such demographic data as age, educational level, and position in the organization. This information helps us describe the sample and enable us to see later if demographic characteristics make a difference in the results. For example, we might discover during data analysis that older employees sense less discrimination than their younger counterparts. Such information could even provide a basis for further theory development. Data Analysis Having collected the data, the researcher must next analyze them, using statistical procedures, to test whether the hypotheses have been substantiated. In the case of the discrimination hypothesis, if a correlational analysis between the variables of mentoring and discrimination indicates a significant negative correlation, the hypothesis will have been supported; that is, we have been correct in conjecturing that the more the supervisor is perceived as a mentor, the less the employee feels discriminated against. Each of the hypotheses formulated from the theoretical framework is tested, and the results are examined. Interpreting Results and Drawing Conclusions The final step is to interpret the results of the data analysis and draw conclusions about them. In our example, if a significant negative relationship is indeed found between mentoring and discrimination, then one of our conclusions might be that mentoring helps fight feelings of discrimination. We might therefore recommend that if the organization wants to create a climate where employees do not feel discriminated against, supervisors should actively engage in mentoring. If the organization accepts this recommendation, it might conduct training programs to make supervisors better mentors. By testing and substantiating each of the hypotheses, we might find a multitude of solutions to overcome the perception of discrimination by employees. Summary We can see that every step in the research process is important. Unless the problem is well defined, the research endeavor will be fruitless. If a thorough literature survey is not done, a defensible theoretical framework cannot be developed and useful hypotheses cannot be generated—which compromises effective problem solving. Using the correct methods in data gathering and analysis and drawing relevant conclusions are all indispensable methodological steps for conducting empirical research. We next examine some of the research design issues which are integral to conducting good research. RESEARCH DESIGN Issues regarding research design relate particularly to how the variables are measured, how the data are collected, what sampling design is used, and how the data are analyzed. Before decisions in these areas are made, some details about the nature and purpose of the study have to be determined so there is a good match between the purpose of the study and the design choices. If the research design does not mesh with the research goals, the right solutions will not be found. Important Concepts in Research Design Five important concepts in research design must be understood before an adequate design can be created: nature of study, study setting, types of study, researcher interference, and time horizon. The nature of study is the purpose of the study—whether it is to establish correlations among variables or causation. The study setting could be either the environment in which the phenomena studied normally and naturally occur—the field—or it could be in a contrived, artificial setting—the laboratory. The type of study is either experimental (to establish causal connections) or correlational (to establish correlations). Experiments can be conducted in an artificial setting—a lab experiment, or it could be conducted in the organization itself where events naturally occur—field experiment. Researcher interference is the extent to which the researcher manipulates the independent variable and controls other contaminating factors in the study setting that are likely to affect the cause–effect relationship. The time horizon is the number of data collection points in the study; the study could be either one-shot (various types of data are collected only once during the investigation) or longitudinal (same or similar data are collected more than once from the same system during the course of the study). Purpose of Study and Design Choices One of the primary issues to consider before making any research design decision is the purpose of the study. Is the research to establish a causal relationship (that variable X causes variable Y), or is it to detect any correlations that might exist between two or more variables? A study to establish a cause–effect relationship differs in many areas (for example, the setting, type of study, extent of researcher interference with the ongoing processes, and time frame of the study) from a study to examine correlations among factors. Figure B.2 depicts the fit between the goal of the study and the characteristics of the study. Causal Studies Studies conducted to detect causal relationships call for an experimental design, considerable researcher interference, and a longitudinal time span. The design could consist of laboratory experiments, field experiments, or simulations. LABORATORY EXPERIMENTS A rigorous causal study may call for a laboratory experiment, where participants are exposed to an artificial environment and an artificial stimulus in order to establish a cause–effect relationship. The experiment is set up with maximum researcher interference; both manipulation and controls (described later) are used, and data are collected from the subjects more than once during the experiment (longitudinally). Following is an example of how a lab experiment is conducted. Suppose a manager wants to know which of two incentives—offering stock options or giving a bonus— would better improve employee productivity. To determine this, the manager has to experiment with each of the two types of incentives to see which offers better results. Not knowing how to proceed, the manager might hire a researcher, who is likely to recommend conducting a lab experiment first, and then a field experiment. The lab experiment firmly establishes the causal relationship, and the field experiment confirms whether or not the causal relationship established during the lab experiment holds good in the organizational setting. To set up a lab experiment in which thirty subjects participate, the following is done: 1. An artificial setting is created. It will consist of three conference rooms in which the experiment is conducted after regular work hours. 2. A simple job—assembling cardboard houses—is given to the subjects who take part in the experiment for two hours. 3. The subjects receive an imaginary bonus in the form of cardboard chips and stock options in the form of fake certificates. 4. Each subject is randomly assigned to one of the three conference rooms, so three ten-member groups are formed. For the first hour, all three groups will be assigned the task of assembling the cardboard houses. Thereafter, the researcher manipulates the incentives—giving one group stock options; another a bonus; and a third, called the control group, no incentives at all. The researcher has already exercised tight control to ensure that all three groups have more or less the same types of members in terms of ability, experience, and the like by randomly assigning members to each of the groups. In random assignment, every member has an equal chance of being assigned to any of the groups. This control helps avoid contamination of the cause–effect relationship, since all factors that might affect the causal relationship (age, ability, and so on) are randomly distributed among the groups. The data are collected at two different times, in the following manner. At the end of the first hour, when all three groups have worked without any incentives, the number of cardboard houses built by each group will be recorded by the researcher. The numbers are again counted and recorded at the end of the second hour, after the introduction of the incentives. Determining the difference between the number of houses assembled during the second hour and the number assembled during the first hour for the three groups clarifies the following two issues: l l Do the incentives make any difference at all to performance? Obviously, if the performance has increased during the second hour for either or both of the two groups provided with incentives, while there is no difference for the control group, then it is safe to surmise that either or both of the incentives have caused performance to rise. If there is no difference in the production between the second and first hour for all three groups, then, of course, the incentives have not caused an increase in performance. If the incentives do make a difference, which of the two incentives has worked better? By examining which group—the group that received the stock options or the group that received the bonus— performed better during the second hour, we know which of the two incentives worked better. The incentive that increases performance more will obviously be preferred by the company. Because all possible contaminating factors have been controlled by the random assignment of members to the three groups, the cause–effect relationships found can be accepted with a high degree of confidence. FIELD EXPERIMENTS What occurred in the tightly controlled artificial lab setting may or may not happen in an organizational setting, where many of the factors (such as employees’ ages and experience) cannot be controlled and the jobs to be done might be quite complex. But having established a strong causal relationship in the lab setting, the researcher is eager to see if the causal relationship found in the lab setting is generalizable to the organization, or field setting. For the field experiment, three experimental cells (three branches or departments of the company, or whatever other units are appropriate for the organization) can be chosen. Real bonus and stock options can be offered to two groups, while the third group is treated as a control group and given no incentives. Work performance data can be collected for the three cells before the incentives are introduced and again six months after the incentives are introduced. While it is possible to manipulate the incentive in a field experiment, it is not possible to control the contaminating factors (ability, experience, and so on). Because employees are already placed, members cannot be randomly assigned to the three units. Under the circumstances, researcher interference can be only partial, since the independent variable can be manipulated but other factors cannot be controlled. Even manipulating the independent variable is not easy, because people in organizations get suspicious and anxious as the word spreads that some strange changes are being made at some sites. Not only does this cause apprehension among employees, but it may also produce invalid results. Because of these difficulties, very few field experiments are conducted in organizational behavior research. However, if the manipulation is successful and the results of the field experiment are similar to those of the lab experiment, the manager can confidently introduce the changes needed to obtain the desired results. If you read journal articles describing experimental designs, you will want to see how well the manipulations were done (were the independent variables successfully manipulated, or did the subjects see through the manipulations?) and how tightly the contaminating variables were controlled. If the subjects in the various groups differed in some characteristics that are relevant to the cause-effect relationship, then it cannot be said with confidence that only the manipulated independent variable caused the dependent variable. Other factors in the setting might also have influenced the dependent variable, and they might be impossible to trace.1 SIMULATIONS Somewhere between lab and field experiments are simulations—experiments that are conducted in settings that closely resemble field settings. The specially created settings look much like actual environments in which events normally occur—for example, offices with desks, computers, and phones. Members of the experimental group are randomly selected and exposed to real-world experiences over a period of time, during which their behavior is studied. A free simulation for studying leadership styles, called “Looking Glass,” has been used in management classes. This simulation enables students to study different kinds of behavior as the researcher manipulates some of the stimuli while allowing the flow of events to be governed by the reactions of the participants.2 Correlational Studies Researchers and managers may not be interested in establishing causal connections. Instead, they may want to understand, describe, or predict occurrences in the workplace. In general, they want to know which factors are related to desirable outcomes (such as employee loyalty to the organization) and which to undesirable outcomes (such as high turnover rates). Correlational studies are studies that are not specifically geared toward establishing cause–effect relationships. Such studies may be exploratory—trying to understand certain relationships; descriptive—trying to describe certain phenomena at the workplace; or analytical—focusing on testing hypotheses. Correlational studies are always conducted in the field setting with minimum researcher interference, and they can be either one-shot or longitudinal. The vast majority of the research articles published in organizational behavior journals are field studies examining correlations among factors. To conduct a scientific study, whether causal or correlational, certain research design decisions must be made. As Figure B.3 shows, these decisions involve measurement, issues, data collection methods, sampling design, and data analysis procedures. Measurement Issues We saw earlier that it is difficult to measure attitudes, feelings, and other abstract concepts. Since the measurement of variables in the organizational sciences is not as exact as in the physical sciences, management research cannot be completely scientific. It is possible, however, to minimize biases in measurement by carefully developing valid and reliable measures for even abstract concepts. The primary aspects in measurement are: operational definition, the “goodness” of measures, and the measurement techniques to be used. OPERATIONAL DEFINITION Attitudes such as job satisfaction and organizational commitment do not easily lend themselves to measurement. To measure them, we first need to translate them into observable behaviors. Operational definition is the reduction of the level of abstraction of concepts so as to identify the observable behaviors and measure them. For example, how can we measure the motivational level of individuals? We know that highly motivated people engage in the following types of behaviors, among others: They are driven by work, and they keep persevering even if they initially fail to accomplish what they want. We can measure the abstract concept of motivation by developing an instrument that asks subjects to respond to several suitably worded questions tapping these behaviors.3 Most of the abstract concepts that are important to the study of organizational behavior have been operationally defined by scholars, who have developed “good” instruments for measuring them.4 “GOODNESS” OF MEASURES “Good” measurement instruments offer researchers the confidence that they do indeed measure what is desired to be measured and in a thorough and accurate manner. The goodness of instruments is established through their validity and reliability. Validity is our confidence that the instrument used does indeed measure the concept it is supposed to measure. For instance, if a twenty-item instrument is developed to measure job satisfaction, we need to know that it does indeed measure job satisfaction, not employees’ general happiness. Researchers usually establish various types of validity for the measures they use. Among them are content validity, criterion-related validity, predictive validity, construct validity, and convergent and discriminant validity. Journal articles often explain the types of validity established for the instrument used, especially if it is newly developed. In general, only such measures as are both valid and reliable are frequently used by researchers.5 Reliability is the ability of an instrument to accurately and stably measure a concept over time and across situations. For example, it is not enough for an instrument to measure job satisfaction; it must do so consistently and accurately time and again in all settings. Most researchers discuss the reliability of their instruments in terms of stability and consistency. Test–retest reliability is one indicator of the stability of a measure over time. Cronbach’s alpha and split-half reliability are two indicators of the internal consistency of instruments. These are the terms you are likely to come across in published empirical research. Authors of studies usually provide details of the measures they use, and at a minimum cite their source. Journal editors and reviewers try to ensure that studies to be published have used valid and reliable measures. Discriminating readers of journals reporting empirical studies pay attention to the “goodness” of the measures. If variables are not validly and reliably measured, how can we place any confidence in the results of the study? MEASUREMENT TECHNIQUES Concepts are not measured solely through questionnaires or interviews. Sometimes, in order to tap certain ideas, feelings, and thoughts that are not easily verbalized, researchers use projective tests—word association, sentence completion, thematic apperception tests, and ink-blot tests are some familiar projective tests. In word association (e.g., Work could be associated with excitement or drudgery,) and sentence completion (“I Like ”) tests, the respondent is believed to give answers based on deeply embedded feelings, attitudes, and orientations. Marketing researchers use these techniques to assess consumer preferences. Thematic apperception tests and ink-blot tests ask the subject to offer a story or interpret an ink blot. They can be interpreted only by trained psychologists. Data Collection Methods Data can be collected through questionnaires, interviews, computers, observation, unobtrusive methods, or a combination of these. The most frequently used method in organizational behavior research is questionnaires. QUESTIONNAIRES A questionnaire is a written set of questions to which respondents record their answers, usually within a close range of alternatives given to them. Questionnaires can be mailed to respondents or administered personally. Mail questionnaires are commonly used because of the large number of people who can be reached economically even when they are geographically dispersed. As a rule, however, they do not elicit a good response rate, even when stamped, self-addressed envelopes are enclosed for their return. (Researchers sometimes even include, as a small token of their appreciation, a $1 bill.) A 30 percent response rate for mail questionnaires is considered good. Mail responses generally fall far short of even this low percentage. Because of the low response rate, certain types of nonresponse biases can creep into research. For example, we cannot know if those who responded to the survey differ from those who did not. Thus, we cannot be sure that the data are representative of the population we are trying to study. Personally administered questionnaires are questionnaires given to groups of subjects by the researcher, who collects the responses immediately after completion. This method ensures practically a 100 percent response rate. However, many organizations are reluctant to spare company time for the research effort unless the study is of vital importance to them. INTERVIEWS Interviews have the potential to elicit a good deal of information. In structured interviews, specific questions are asked of all respondents, and the responses are noted down by the interviewer. In unstructured interviews, there is no predetermined format; questions are framed according to responses given to the previous question. Structured interviews are conducted when the interviewer knows precisely what sort of information is needed. They are efficient in terms of the amount of time involved in both obtaining the required information and categorizing the data obtained. Unstructured interviews are conducted when the researcher wants to explore a problem or become more knowledgeable about particular situations. Face-to-face interviews offer the researcher the advantage of being able to observe the interviewees as they respond to questions. Nonverbal messages transmitted by the interviewees can be observed and explored further. Telephone interviews, on the other hand, help the researcher reach a vast number of geographically dispersed individuals. In both face-to-face and telephone interviews, certain types of biases can enter. The way a question is worded and asked, the inflection of a voice, the frame of mind of the interviewee at the time the interview is conducted, and other factors can all contribute to biases in the data. COMPUTERS Computer-assisted interviewing and computer-aided surveys will become more popular in the future as more and more people become comfortable using their computers at home and responding to questions contained on diskettes or displayed on web sites. Interview and questionnaire methods of data collection are greatly facilitated through computers. However, computer literacy of respondents is prerequisite for using computer-assisted data collection techniques effectively. OBSERVATIONAL SURVEYS Observational surveys are another data collection method whereby information is obtained without asking questions of subjects. In this method, the researcher observes firsthand what is going on and how people are behaving in the work setting. The data are collected by either nonparticipant observers (researchers who observe behavior as outsiders) or participant observers (integral members of the work team). An example of a nonparticipant study is one done by Henry Mintzberg, who observed the nature of managerial work over a period of time. Like interviews, observational surveys can be either structured or unstructured. In a structured observational survey, the observer identifies the factors that are to be observed. For example, the observer might want to note the number of times a manager gives instructions to staff members and how much time this takes. In an unstructured observational survey, the observer might simply want to know how the manager spends the day at the workplace and might jot down all the activities the manager engages in and the time periods and frequencies involved. Observational studies help prevent respondent bias, since information is not given by the subjects directly. Any bias that might creep in through the self-consciousness of subjects usually lasts only a few days. Then, subjects begin to function and behave normally, oblivious to the presence of the observer. However, observer fatigue and observer bias cannot be totally avoided in observational studies. Moreover, when several observers are involved in a large research project, interobserver reliability could become an issue for concern; different observers might interpret and categorize the same behavior differently. This problem can be minimized by training the observers before the start of the project. UNOBTRUSIVE METHODS Data collection by unobtrusive methods offers valid and reliable information; bias is minimized because the source of the data is tangible elements rather than people. For example, the usage of library books can be determined by the wear and tear on them, a source of information more reliable than surveys of users of the library. The number of cans or bottles of beer in the recycling bins outside houses on garbage collection days would offer a good idea of the beer consumption patterns in households. The personnel records of a company would indicate the absenteeism patterns of employees. Unobtrusive methods thus have the potential to offer the most reliable and unbiased data. They are, however, time consuming and labor intensive; also, the researcher must obtain the company’s permission to gain access to such data. MULTIPLE METHODS Each data collection method has advantages and disadvantages. The best approach is using multiple methods of collecting data, since it offers researchers a chance to cross-check the information obtained through the various methods. This approach, however, is expensive and hence is used infrequently in organizational behavior research. When you read journal articles, you should assess the data collection methods used by the researchers to determine if they are adequate. Authors of published studies often discuss the limitations of their research and the biases they have attempted to minimize. The biases could relate to the types of measures used, the data collection methods adopted, the sampling design, and other research process and design issues. Sophisticated managers pay attention to all research design details in order to evaluate the quality of the research. Sampling Design Sampling is the process of drawing a limited number of subjects from a larger population, or universe. Since researchers cannot possibly survey the entire universe of people they are interested in studying, they usually draw a sample of subjects from the population for investigation. The sampling design used makes a difference in the generalizability of the findings and determines the usefulness and scientific nature of the study. Sample size is another important issue. There are two broad categories of sampling—probability sampling and nonprobability sampling. PROBABILITY SAMPLING Probability sampling is sampling that ensures that the elements in the population have some known chance, or probability, of being selected for the sample. Because of this, probability sampling designs offer more generalizability than nonprobability designs. There are many probability designs. The simple random sampling design, wherein every element in the population has a known and equal chance of being chosen, lends itself to the greatest generalizability. However, other probability designs can be more efficient and offer good generalizability as well. Among them are systematic sampling, stratified random sampling, cluster sampling, and area sampling. In systematic sampling, every nth element in the population is chosen as a subject. In stratified random sampling, the population is first divided into meaningful strata (for example, blue-collar and white-collar employees); a sample is then drawn from each stratum using either simple random sampling or systematic sampling. Cluster sampling is the random selection of chunks (clusters or groups) of elements from the population; every chunk has an equal chance of being selected, and all the members in each chosen chunk participate in the research. For example, in an attitude survey, three departments in an organization can be randomly chosen; all the members of the three departments are the subjects. Area sampling is cluster sampling confined to particular geographical areas, such as counties or city blocks. Marketing researchers use cluster and area sampling extensively for surveys. NONPROBABILITY SAMPLING For some research projects, probability sampling may be impossible or inappropriate. In such cases, nonprobability sampling may be used, even if generalizability is impaired or lost. In nonprobability sampling, the subjects do not have a known probability of being chosen for the study. For instance, the sample of subjects in a study of sexual harassment must come from those who have experienced such harassment; there is nothing to be gained by researching all the employees of the organization. When the choice of subjects for a study involves a limited number of people who are in a position to provide the required information, a probability sampling design is infeasible. The results of such a study are not generalizable; nevertheless, this type of sampling is the best way to learn about certain problems, such as sexual harassment. Nonprobability sampling includes convenience sampling, judgment sampling, and quota sampling. In convenience sampling, information is collected from whoever is conveniently available. In judgment sampling, subjects who are in the best position to provide the required information are chosen. In quota sampling, people from different groups—some of which are underrepresented—are sampled for comparison purposes. One example might be a study of middle-class African-Americans and whites. As noted earlier, nonprobability sampling does not lend itself to generalizability. In reading research articles, you should determine the type of sampling design being used and how much generalizability the author claims for the research. SAMPLE SIZE Another critical issue in sampling is sample size. Too small or too large a sample could distort the results of the research. Tables providing ideal sample sizes for desired levels of precision and confidence are available to researchers. In examining any business report or journal article, you should note the sampling design and the sample size used by the researcher to assess the generalizability of the findings. Data Analysis Procedures Beyond good measures, appropriate data collection methods, and an acceptable sampling design, a good research project should also have suitable data analysis procedures. Some data cannot be subjected to sophisticated statistical tests. Two examples are data collected on a nominal scale, that divides subjects into mutually exclusive groups, such as men and women or the poor and the rich; and data collected on an ordinal scale, that rank-orders the subjects and indicates a preference (X is better than Y). Various simple ways are available to analyze such data that are qualitative or nonparametric in nature. For instance, if we have categorized under distinct heading the verbal responses of organizational members to an open-ended question on how they perceive their work environment, a frequency count of the responses in each category would be adequate to describe how the work environment is perceived. Likewise, to detect if the gender of the worker (male versus female) is independent of members’ commitment to the organization (less committed versus more committed) a simple x2 (chi-square) test would suffice. Sophisticated statistical tests are possible when data have been gathered on interval or ratio scales. Data collected on interval scales—through individuals’ responses to questions on equal-appearing multipoint scales— allow for the computation of the arithmetic mean and standard deviation. Data collected on ratio scales also allow us to compute proportions and ratios. For example, an individual who weighs 250 pounds is twice as heavy as one who weighs 125 pounds. Pearson correlations can be calculated, and multiple regression and many multivariate analyses can be made with data obtained on interval and ratio scales. These sorts of analyses cannot be made with data obtained on nominal and ratio scales. Illustrations of the four scales appear in Figure B.4. One decision that needs to be made before collecting the data is: What kinds of analyses are needed to find answers to the research question? This decision will determine which scales should be used in data collection. Sometimes researchers are tempted to apply more sophisticated statistical analyses to data that do not lend themselves to such analyses (this includes sample sizes below thirty). Using inappropriate methods can negatively affect the interpretation of the results and can compromise the problem solution. Biases in Interpretation of Results Thus far, we have examined the biases that would result from poor research process and design decisions. Another source of bias is in the interpretation of results. Objectivity plays a large part in the validity of interpretations from the results of data analysis. Objectivity may be difficult, however, if the results of the study do not substantiate the theories painstakingly developed by the researcher. When data analysis does not substantiate one or more of the hypotheses generated, the researcher may be tempted to downplay the results or try to explain them away. For example, a researcher may say that the results were actually in the expected direction even though they were not statistically significant. If a hypothesis has not passed the appropriate statistical test, the hypothesis is just not substantiated, regardless of whether the results were in the theorized direction. When authors try to explain their results, you have to decide for yourself whether the explanations offered are valid. ORGANIZATIONAL BEHAVIOR RESEARCH AND YOU It is seldom possible to do completely scientific research in the field of organizational behavior. First, adherence to good research design principles may not always be possible, since certain choices (such as obtaining the most representative sample for better generalizability or utilizing the best data collection methods) may be beyond the researcher’s control. Second, attitudes and feelings cannot be measured accurately. Hence there are likely to be several types of biases in research in this field. However, by paying careful attention to the research process and rigorously making good research design choices, we are able to minimize the biases and enhance the objectivity, testability, replicability, precision and confidence, and generalizability of our research. Bias can enter at every stage of the process, from problem definition to problem solution. Errors can creep into experimental designs by way of poor or inadequate manipulations and controls. They can enter into measurement, data collection, sampling, data analysis, interpretation of results, and the drawing of conclusions therefrom. Unless managers are knowledgeable about some of the methodological flaws that can adversely affect research results, they may inappropriately apply the conclusions drawn in published research to their own settings. Having been exposed to the rudiments of scientific research, you can critically examine and evaluate all published works before you assess their usefulness for your organization. For instance, you would not consider applying the results of good research done in a service organization to a manufacturing firm. Good research results in the hands of knowledgeable managers are highly useful tools. That is where research knowledge becomes invaluable. By grasping the essentials of good research, you will become a discriminating consumer of business reports and published articles and can become an effective manager. Research knowledge can often make the difference between managerial excellence and mediocrity. References 1. Two sources for further reference on experimental design are D. T. Campbell and J. C. Stanley, Experimental and Quasi-Experimental Designs for Research (Chicago: Rand McNally, 1966); and T. D. Cook and D. T. Campbell, Quasi-Experimentation: Design and Analysis Issues for Field Settings (Boston: Houghton Mifflin, 1979). 2. M. L. Lombardo, M. McCall, and D. L. DeVries, Looking Glass (Glenview, Ill.: Scott, Foresman, 1983). 3. Elaboration of how such measures are developed is beyond the scope of this appendix but can be found in U. Sekaran, Research Methods for Business: A Skill Building Approach, 2nd ed. (New York: Wiley, 1992). 4. Several measures are available in Psychological Measurement Year-books; J. L. Price, Handbook of Organizational Measurement (Lexington, Mass.: D. C. Heath, 1972); and Michigan Organizational Assessment Packages (Ann Arbor, Mich.: Institute of Survey Research). 5. One such instrument is the Job Descriptive Index, which is used to measure job satisfaction. It was developed by P. C. Smith, L. Kendall, and C. Hulin. See their book The Measurement of Satisfaction in Work and Retirement (Chicago: Rand McNally, 1969), pp. 79–84. F I G U R E B . 1 Steps in the Hypothetico-Deductive Approach to Research F I G U R E B . 2 Fit between Goal of Study and Study Characteristics F I G U R E B . 3 Research Design Decisions F I G U R E B . 4 Illustrations of Four Data Analysis Scales