PsyQuesta Learn Psychology with Afa PSYCHOLOGICAL MEASUREMENT AND TESTING B.Sc Psychology SEMESTER -3 PsyQuesta Learn Psychology with Afa MODULE- 1 INTRODUCTION TO MEASUREMENT AND SCALING TECHNIQUES A. DEFINITION OF MEASUREMENT “Assignments of numerals according to rules.” - Tyler “Measurement consists of rules for assigning numbers to object in such a way as to represent quantities of attributes.” - Nunnally Measurement is the assignment of scores of individuals so that the scores represent some characteristics of the individuals. Psychological measurement can be achieved in a wide variety of ways, including self- report, behavioural and physiological measures. B. SCALES OF MEASUREMENT It is the way that numbers are assigned determine scale of measurement. Each scale of measurement represents a particular property or set of properties of the abstract number system. C. LEVELS OF MEASUREMENTS 1. Nominal scale Lowest level of measurement Numbers are used to name, identify or classify persons, objects or groups Nominal scales are really not scales and their only purpose is to name objects Here, members of any two groups are never equivalent but all members of any one group are always equivalent. Mathematical operations like addition, subtraction, multiplication and division are not possible Admissible statistical operations: counting or frequency, percentage, proportion, mode and coefficient of contingency. Drawback: this scale is most elementary and simple 2. Ordinal scale Second level of measurement where there is property of magnitude but not of equal interval or an absolute zero. Numbers denote the rank order of the objects or individuals. Here numbers are arranged from highest to lowest or lowest to highest. This measure reflects which person is larger or smaller, heavier or lighter, brighter or duller etc. PsyQuesta Learn Psychology with Afa E.g., socioeconomic status of people. Every member of upper class is higher in social prestige than every member of middle or lower classes. Likewise, every member of middle class is higher in social prestige than that of lower class. Permissible statistical operations: median, percentile, rank correlation coefficient + all those are permissible for nominal scale. Drawback: ordinal measures are not absolute quantities nor do they convey that the distance between the different rank values are equal. This is because they are not equal interval measurements nor do they incorporate absolute zero. there is no way to ascertain whether a person has any of the characteristics being measured. 3. Interval scale Third level of measurement Includes all the characteristics of nominal and ordinal scale The salient feature of this scale that numerically equal distances in the scale indicate equal distances in the properties of the objects being measured. Or we can say the unit of measurement is constant and equal This scale does not have an absolute zero rather than the zero point which is arbitrary, i.e., the zero point does not tell the real absence of the property being measured. Permissible statistical operations: AM (arithmetic mean), SD (standard deviation), Pearson R, t-test, F-test and other statistics based upon them. Drawback: coefficient of variation cannot be applied. 4. Ratio scale Highest level of measurement Has all the properties of nominal, ordinal and interval scales plus absolute zero point Salient feature of this scale: ratio of any two numbers is independent of the unit of measurement Anything that can be measured from absolute zero can be measured within a ratio scale. E.g., measures of high, weight, length, width etc. Here, the zero point is not arbitrary, but true. All statistical operations including coefficient of variation can be used. In social sciences like psychology and sociology, we frequently encounter with interval measurement because most of the data obtained from measurement are such that we assume an equal unit of measurement and arbitrary zero point. PsyQuesta Learn Psychology with Afa D. PROPERTIES OF LEVELS OF MEASUREMENT 1. Identity Assignment of numbers to respondent’s response and these numbers are just for the sake of identification. These numbers can’t be used for mathematical operations. E.g., roll number of students in a class. 2. Magnitude Property of “moreness”. Any scale is said to have the property of magnitude if it can be said that a particular instance of the attributes represents more, less or equal amounts of the given quantity that does another instance. E.g., height and weight of students in a class. 3. Equal interval A scale has the property of equal interval if the difference between any two points at any place on the scale has the same meaning as the difference between two other points that differ by the same number of scale units. E.g., difference between 4kg and 6kg on a weight measuring scale represent the same quantity as the difference between 14 kg and 16 kg, i.e., exactly 2 kg. Psychological tests rarely have the property of equal interval. For e.g., the difference between IQ 50 & 60 does not mean the difference between IQ 110 & 120, although each of these differ by 10 points. 4. Absolute zero An absolute zero is said to exist when nothing of the property being measured exist (temperature on 0 is not absolute zero as it still has some effect and we can’t say no temperature). It is very difficult or impossible for a psychological measurement to define zero. PsyQuesta Learn Psychology with Afa DISTINCTION BETWEEN PSYCHOLOGICAL MEASUREMENT AND PHYSICAL MEASUREMENT. PHYSICAL MEASUREMENT PSYCHOLOGICAL MEASUREMENT 1. Unit of measurement is fixedand constant throughout the measurement 1. Unit of measurement not fixed and varies throughout the measurement 2. There is a true zero point 2. There is an arbitrary zero point 3. More accurate and predictable 3. Less accurate and predicable 4. Measurement is direct 5. Entire quantity can be measured 4. Measurement is indirect 5. Entire quantity cannot be measure PsyQuesta Learn Psychology with Afa Ps ycho logic al (or qualitative) measurement comprises the measurement of mental processes, traits, habits, tendencies and likes of an individual. Physical (or quantitative) measurement comprises the measurement of objects,height, weight, length etc. which are physically present in the world E. PROBLEMS IN PSYCHOLOGICAL MEASUREMENT 1. Indirectness of measurement Most psychological and educational measurement are indirect. This is because psychological and educational variables cannot be observed and studied directly. E.g., intelligence of student cannot be seen, touched or experienced. 2. Incompleteness of measurement Psychological measures are generally incomplete and therefore measurement of psychological variables are also incomplete. E.g., when an investigating is assessing the attitude towards coeducation, he is required to construct a scale in which a number of samples of behaviour expressing such an attitude need to be incorporated. This number has no limit. Any attempt, to measure such an attitude would be partial and incomplete. 3. Relativity in measurement Psychological measurements are relative. This can be explained with an example: suppose Mohan, a student of class X was given arithmetical knowledge and English language test. Let us further suppose he scored 95% in English language but could not answer even a single item on the arithmetic test. We cannot say Mohan is a brilliant student on the basis English test result nor we can say he is dull on the basis of arithmetic test result. 4. Errors in measurement Measurement in physical sciences as well as behavioural sciences is most of the time not pure. It contains some uncontrolled factors, which produce some gross errors. Sources of errors in measurement are given below. Sources of errors in measurement: 1) Respondent: - One of important sources of errors is the respondent himself. Sometimes, the respondent is found reluctant to express his/her true feelings or it PsyQuesta Learn Psychology with Afa may be that sometimes due to lack of knowledge the person may not express himself clearly. In either case, the measurement loses its accuracy 2) Measurer: -Sometimes the behaviour, style and looks of the person who is measuring the phenomena, distort the process of measurement. His behaviour, style and looks may encourage or discourage certain types of replies from the respondent that affects the accuracy of the measurement. 3) Situation: - Situational factors also contribute to errors in measurement. Any situation that puts unnecessary strain on the respondent, tends to introduce errors in measurement. 4) Test instruments: -Poor psychometric qualities of the test and defective measuring instrument contribute to errors in measurement. F. CONCEPTS OF PSYCHOPHYSICS Psychophysics is defined as the scientific study of relationship between stimuli and sensations & perceptions evoked by these stimuli. It was introduced by Gustav Theoder Fechner in 1860’s. 1. Psychophysical methods: - A set of procedures that have developed to investigate sensory thresholds 2. Absolute threshold: - Absolute threshold is the lowest level of stimulus (light, sound, touch etc.) which an organism can detect at least 50% of the time. It is also known as absolute limen (AL), detection threshold, sensation threshold etc. 3. Difference threshold: - Difference threshold is the smallest amount of change in physical stimulus necessary for an individual to notice a difference in the intensity of the stimulus. It is also called difference limen (DL), just noticeable difference (JND) etc. 4. Subliminal perception: -Subliminal perception occurs when a stimulus is too weak to be perceived yet a person is influenced by it. A signal detected 49% of the time is would be subliminal. 5. Weber’s law Developed by EH Weber. It is the first systematic attempt to formulate a principle which governed the relationship between psychological experience and physical stimulus. There are two types of stimuli: Standard stimulus: fixed, no change in intensity or properties Comparable stimulus: varies, intensity or properties can be altered “When the magnitude of standard stimulus is increased, the size of change needed for PsyQuesta Learn Psychology with Afa discrimination between the standard and comparable stimulus (or JND) is also increased.” E.g., candles in a room Equation: ∆R = K where, R ∆R = difference threshold OR just noticeable difference (JND) R = standard stimulus K = constant Constant K in Weber’s Law is always a fraction known as Weber’s fraction or proportion. It indicates the proportion by which the standard stimulus must be increased in order to produce just noticeable difference (JND). Weber’s law is regarded as a good measure of the overall sensitivity in different sense modalities. Drawback: precision is lost when the standard stimulus reaches the extremes, i.e., when the standard stimulus become either very weak or very strong, the precision is lost to a great extent. 6. Fechner’s law Developed by Gustav Theodor Fechner Fechner’s law is an improvement up on Weber’s Law Fechner’s law states that stimulus value and resulting psychological sensations have a logarithmic relationship i.e., when the stimulus value increase in geometrical progression, the psychological sensations increase in arithmetical progression. Equation: R= K log S where, R=magnitude of sensation of response K=Weber’s constant S=magnitude of stimulus value In other words, the law states that the magnitude of sensation (or response) varies directly with the logarithm of the stimulus value Increasement by the process of multiplication is termed as geometrical progression and increasement by the process of addition is called arithmetic progression. E.g., perceived loudness or brightness is proportional to logarithm of the actual intensity measured with an accurate non-human instrument. H.PSYCHOPHYSICAL & PSYCHOLOGICAL SCALING METHODS Scaling methods are methods through which stimuli or individuals are sorted according to some known and specified characteristics or attributes. Depending upon the classes of stimuli, there are two types of scaling methods: psychophysical and psychological scaling method Psychophysical scaling methods In this method, the stimuli presented have a physical referent, i.e., it can be PsyQuesta Learn Psychology with Afa described by a physical scale. E.g., grams, decibels, meter, centimetre etc… The purpose of the psychophysical scaling method is to discover some definite quantitative relationship subjective measurement of stimuli (as judged by the subject) and the objective measurement by physical scales. E.g., Method of average error (or method of adjustment), method of minimal changes (or method of limits), method of constant stimuli (or method of right and wrong cases) etc… Psychological scaling methods: In psychological scaling methods, stimuli possess psychological attributes, which lack an appropriate physical referent. According to Torgerson, “psychological scaling methods are procedures for constructing scales for the measurement of psychological attributes”. For measuring psychological attributes like sense of humour, extend of honesty, aggressiveness, pleasantness and so on, there are no physical scales. Therefore, these attributes are best measured by the psychological scales. 1. Method of average error 2. Method of minimal changes 3. Method of constant stimuli 4. Method of pair comparison 5. Method of rank order Method Of Average Error Also known as method of adjustment, method of reproduction or method of equivalent stimuli Introduced by Gustav Fechner Subject is provided with a Standard Stimulus (St) and Comparable stimulus (Co) which maybe greater or lesser in intensity than standard stimulus. The subject is then required to adjust the Comparable stimulus until it appears to be equivalent to standard stimulus. The difference between St and Co defines the error in each judgement. A large number of such judgments are obtained and the arithmetic mean of those judgments are obtained. Hence the name method of average error The value thus obtained is the value for Point of Subjective Equality (PSE). The difference between St and PSE indicates the presence of the constant error (CE) The purpose of this method is to determine equivalent stimuli by active adjustment of the Co by the subject in each trial and hence, the method is also known as “method of equivalent stimuli”. Method of Minimal Changes It is also known as method of limits, method of JND. For computing threshold in this method, two modes of presentation are PsyQuesta Learn Psychology with Afa adopted: ascending series, descending series For computing differential threshold, the comparable stimulus (Co) is varied in possible small steps either in ascending or descending series and the subject is required to say in each step whether the Co is smaller (-), equal (=) or greater (+) than the St. For computing absolute threshold no St is needed, the subject simply reports if they can detect the change in stimulus presented in ascending and descending series. Method of Constant Stimuli Also known as method of right and wrong cases, method of frequency In this method a number of constant stimuli are presented to the subject several times in random order. To determine absolute threshold, the subject has to report each time whether he perceives or does not perceive the stimulus presented in random order Method of Pair Comparison Formulated by Thurstone in 1927 In this method, stimuli are paired and the subject is required to make a comparative judgement by saying which member of each pair possess most of the traits being scaled. Originally introduced by Cohen in 1894 and was further developed by Thurstone. Extension of method of constant stimuli Each stimulus is compared with every other stimulus and therefore each stimulus serves as a standard stimulus in turn. When the two subject is presented with the two stimuli they are required to tell which one of them is the greater. No equality judgement is allowed. Method of Rank Order Also known as “order of merit”. Developed from works of Cattell and Spearman. In this method, all stimuli are presented simultaneously to the subject who is requested to rank them in order from high to low Rank 1 indicates the best and the higher rank number indicates the inferior position of the stimulus being recorded Most suited when number of stimulus or objects to be ranked is small. No problem regarding the sequences of presentation of object arise. The scale resulting from this method is known as “ordinal scale”. Since objects are assigned higher and lower ranks, there is no way to commit the error of central tendencies. PsyQuesta Learn Psychology with Afa MODULE- 2 NATURE AND USE OF PSYCHOLOGICAL TESTS A.DEFINITION OF PSYCHOLOGICAL TEST It is a standardized instrument consisting of a series of questions called items, which assess certain aspects of a person’s individual ability and describe it in terms of scores and categories. A psychological test is essentially an objective and standardized measure of a sample behaviour.- Anne Anastasia, (1988). B.HISTORICAL BACKGROUND Psychological testing in its modern form originated little 100 years ago in laboratory studies of sensory discrimination and reaction time. It was the English biologist Sir Francis Galton who was primarily responsible for launching the testing movement. A unifying factor in Galton's numerous and varied research activities was his interest in human heredity.He invented the first battery of tests for accessing sensory and motor aspects. He also set up an anthropometric laboratory at the International Exposition of 1884 where, by paying three pence, visitors could be measured in certain physical traits and could take tests of keenness of vision and hearing, muscular strength, reaction time, and other simple sensorimotor functions It was Galton's belief that tests of sensory discrimination could serve as a means of gauging a person's intellect American psychologist James McKeen Cattell Newly established science of experimental psychology and the still newer testing movement merged in Cattell's work. Kraepelin (1895), was interested primarily in the clinical examination of psychiatric patients, he prepared a long series of tests to measure what he regarded as basic factors in the characterization of an individual. The tests employing chiefly simple arithmetic operations, were designed to measure practice effects, memory, and susceptibility to fatique and to distraction. PsyQuesta Learn Psychology with Afa A few years earlier. Oehm (1889), a pupil of Kraepelin, had employed tests of perception, memory, association, and motor functions in an investigation on the interrelations of psychological functions. Another German psychologist Ebbinghaus (1897), administered tests of arithmetic computation memory span, and sentence completion to schoolchildren. The most complex of the three tests, sentence completion, was the only that showed a clear correspondence with the children's scholastic achievement Like Kraepelin, the Italian psychologist Ferrari and his students were interested primarily in the use of tests with pathological cases. An extensive and varied list of tests was proposed, covering such functions as memory imagination attention, comprehension Suggestibility, aesthetic appreciation, and many others. In these tests we can recognize the Brends that were eventually to lead to the development of the famous Binet intelligence Scale. the examination of the mentally ill around the middle of the 19th century resulted in the development of numerous early tests. In 1865 German physician Hubert Von Grashey developed the antecedent of memory drum as a means testing brain injured patients, shortly thereafter German psychiatrist Conrad Riger developed a test battery for brain damage Experimental psychology flourished in the late 1800's in Europe and Great Britain pioneers such as Wundt, Galton, Cattle, Weschler showed that it was possible to expose the mind to scientific scrutiny and measurement. William Wundt founded the first psychological laboratory in 1879. Leipzig in Germany, followed by Galton in Britain. Cattle studied the new experimental psychology with both Wundt and Galton. He invented the term mental test in1819, Thorndike studied under Cattle and made monumental contributions in leaming theory and educational psychology Cattle's other scholar R S Woodworth authored the very popular and influential book Experimental Psychology. Another scholar of Cattle, EK Strong developed a vocational interest, and the most influential scholar of Cattle, Weschler studied extensively on intelligence The first modern intelligence test was invented in 1905 by Alfred Binet. PsyQuesta Learn Psychology with Afa Later in 1908, Binet and Simon published the revised version of the 1905 scale In 1911 a third revision of the Binet Simon scale appeared. Later Stanford-Binet was published in 1916 by incorporating verbal and performance tests The projective approach in psychology was associated with the word asutation mated developed by Francis Chalton Rorschach developed a personality test which is projective in nature known as the Rorschach inkblottast. In 1935 Morgan Murray developed Thematic Apperception Test (TAT) to study personality. In 1928 Payme developed a sentence completion test C. CHARACTERISTICS OF A GOOD TEST I. OBJECTIVITY A test must be free from subjective element so that there is complete interpersonal agreement among experts regarding the meaning of the items and scoring of the test. Objectivity here, relates to two aspects of the test - objectivity of the items and objectivity of the scoring system. By objectivity of items is meant that the items should be phrased in such a manner that they are interpreted in exactly the same way by all those who take the test. For ensuring objectivity of items, item must have uniformity of order of presentation (i.e, either ascending or descending order) By objectivity of scoring is meant that the scoring method of the test should be a standard one so that complete uniformity can be maintained when the test is scored by different experts at different times. II. RELATABILITY A test must also be reliable. Relatability refers to the self correction of the test. It shows the extend to which the result obtained are consistent when the test is administrated once or more than once on the same sample with a reasonable time gap. Consistency in a results is obtained in a single administration is the index of internal consistency of the test and consistency in results obtained upon testing and retesting is an index of temporal consistency Reliability thus includes both internal consistency as well as temporal consistency For a test to be called sound it must be reliable because reliability indicates the extend to which the scores obtained in the test are free from such internal defects of standardization which are likely to produce errors of measurement. PsyQuesta Learn Psychology with Afa III. VALIDITY Validity indicates the extend to which the test measure what it intends to measure, when compared with some outside independent criterion In other words it is the correlation of the test with some outside criterion. Generally, validity of the test depend upon the reliability because a test which yields inconsistent results (poor reliability) is ordinary not expected to correlate with some outside independent criterion. IV. NORMS A test must also be guided by certain norms. Norms refer to the average performance of a representative sample on a given test. There are for common types of norms - age norms, grade norms, percentile norms and standard score norms Depending upon the purpose and use, a test constructor prepares any of these norms for his test. Norms help in interpretation of scores, in the absence of norms no meaning can be added to the score obtained on the test. V. PRACTICABILITY A test must also be practicable from the point of view of the time taken in its completion, length, scoring etc. In other words the test should not be lengthy and scoring method must not be difficult nor one which can only be done by highly specialized persons. D. ETHICAL ISSUES IN PSYCHOLOGICAL TESTING Psychological testing refers to all the possible uses, applications and underlying important concepts of psychological and educational tests. To maintain its proper uses ans applications, the American Psychological Association (APA) has officially adopted a set of standards and rules in 1953 which have undergo continual review and refinement.The main ethical and moral issues relating to psychological testing can be described under the five heading I. ISSUES OF HUMAN RIGHT Various types of human rights have been recognized in the psychological testing. Among these rights is, right to not to be tested. In fact, persons who don’t want to subject themselves to testing, should not and ethically can’t be forced to accept this. Moreover, individuals who finally decide to subject themselves to testing, have rights to know their test scores, their interpretations as well as basis of any decisions that affect their lives. Likewise, these days other human rights such as the right to who will have the access to data of psychological testing, and the right to confidentiality of test results are being popularly discussed. PsyQuesta Learn Psychology with Afa II. ISSUE OF LABELLING On the basis of psychological testing, a person is given a certain label or diagnosed ha having a certain psychiatric disorder. This labeling has many harmful effects Labeling can stigmatize a person for life and it also affects one’s access to help. Such a labeling creates additional problems It also lower tolerance for stress and make treatment difficult In view of these potential negative effect and dangers of labeling a person should have right to not to be labeled. III. ISSUES OF INVASION OF PRIVACY When people responds to items of psychological tests, they have little idea of what is being revealed by their responses, but somehow they feel that their privacy has been invaded. psychological tests have very limited and pinpointed aim and they can’t invade the privacy of the person. psychologists don’t consider it wrong, evil or detrimental to find out or collect information about the person, often the subjects privacy is invaded when information is misused Psychologists are legally and ethically bound to maintain confidentiality and don’t reveal any more information than that is necessary to accomplish the purpose of testing. Ethical code of APA(1992) has included confidentiality, i.e., which obviously indicates that personal information obtained by psychologists from any source is communicated to any others only with their consent. Exception to this exist only if holding the information pose danger for them, others . IV. ISSUES OF DIVIDED LOYALITIES Psychologists face conflicts when individuals welfare is is put at odds on hand and that of the institute that employed psychologist n the other. Psychologist is to maintain test security, but also not violate the person’s right to know the basis of an adverse decision, this decisions when explained to a person might out go to others with same problems, who rightly can decide to outsmart the test V. RESPONSIBILITY OF TEST CONSTRUCTORS AND TEST USERS The test constructor is responsible for providing all the necessary information. Test constructors must provide a test manual which may clearly state the appropriate uses of the test, including data relating to reliability, validity and norms clearly specify about the scoring and administration standards. Test users should have adequate knowledge. He should be aware of psychometric qualities of the test being used as well as the relevant literature At any cost, a test user can not claim ignorance. PsyQuesta Learn Psychology with Afa E. FACTORS INFLUENCING TEST ADMINISTRATION Any influences that are specific to the test situation constitute error variance and reduce test validity. It is therefore important to identify any test related influences that may limit or impair the generalizability of test result. 1. EXAMINER INFLUENCES Memorizing the exact verbal instructions is essential in most individual test. some previous formalities with the statement to be read prevents misleading and hesitation and permits a more natural, informal manner during test administration. In individual test especially in the administration of performance test, such preparation involves the actual layout of necessary materials to facilitate subsequent use with a minimum of search or fun bling. Materials should generally be placed on a table near the testing table so that they are within easy reach of the examiner but do not distract the test taker. In group testing, all test blanks, answer sheets, special pencils or other materials needed should be carefully counted, checked and arranged in advance of the testing day. Whether the examiner is a stranger or someone familiar to the test takers might make a significant difference in scores. Children are more susceptible to examiner and situational influences than are adults. 2. TESTING CONDITION The environment for testing should also be appropriate. Noise-free room, proper seating arrangement, and adequate light should be provided to test takers It is important to recognize the conditions which might affect the test scores e.g noise, privacy, traffic in the testing room etc, even the type of answer sheet employed may affect test scores. Many subtle testing conditions affect the performance on ability as well as on performance tests. 3. TEST TAKER The motivation of the test taker to take a test can greatly influence the test scores. Test takers emotion, fatigue, any sort of handicap, mental set etc influence the test results. It is important that the test taker be relaxed and a good rapport should be established. practices designed to enhance rapport serve also to reduce test anxiety. Procedures tending to dispel surprise and strangeness from the testing situation and to reassure and encourage the subject should certainly help to PsyQuesta Learn Psychology with Afa lower anxiety 4. TEST ADMINISTRATION PROCESS The main purpose of testing is to generalize the results obtained from the sample to those in non-test situations. Any influences that are specific to the test situation constitute error variance and reduce test validity. The testing process itself, examiner related variables, and the subject/examinee related variables all are potential confounding variables that need to be looked into, watched, and thoroughly controlled for the generalizability of the test. F. CLASSIFICATION OF PSYCHOLOGICAL TEST I. SPEED AND POWER TEST SPEED TEST A pure speed test is one in which individual differences, differ in speed of performance Items are arranged in uniformly low difficulty; the ability of test takers are concerned Speed tests trying to make every question correct if they have enough time. E.g: Clinical Speed and Accuracy Test POWER TEST Pure power tests are measure in performance of how much the testtakers knows or can do Time is up to the test. The difficulty of items is steeply graded, and items have a higher difficulty level also,that cant be answered. Power test is one where all test-takers have enough time to do their best; the only concern is what they can do. E.g: Raven’s Progressive Matrices II. INDIVIDUAL TESTS AND GROUP TESTS INDIVIDUAL TEST Administered to only person at a time Individual tests require one examiner and one subject E.g: Kaufman Scale Tests, KohsBlock Design Test. Rapport is essential Examiner’s role is important GROUP TEST Taken by a number of people together at same time One examiner works with many subject together E.g: WAIS, WISC, Standford -Binet Scales, paper and pencil tests Easy to administer, and easy to score. Examiners role is not as important PsyQuesta Learn Psychology with Afa III. VERBAL AND NON VERBAL TEST VERBAL TESTS use language to ask questions and demonstrate answers PERFORMANCE/ NON VERBAL Minimal or no use of language in solving the problems There is no need for a subject to manipulate Subject requires manipulation of objects, sketching maze, arranging pictures and completing patterns Both literate and illiterate can take the test Visual spatial tasks are used, non-verbal responses are yielded Limited to literate people Requires subject to give verbal responses, speaking or writing IV. OBJECTIVE AND SUBJECTIVE OBJECTIVE Test that measures an individual’s characteristics in a way that isn’t influenced by the examiner’s own beliefs; in this way, they are said to be independent of rater bias Objective items include multiplechoice, true-false, matching and completion Reliable and valid E.g: Self- Report Measure SUBJECTIVE An assessment tool that is scored according to or to standards that are less systematic Subjective items include short answer essays, problem solving and performance test items Reliability and validity are less accurate E.g: Case History and Interview V. CULTURE SPECIFIC AND SUBJECTIVE CULTURE SPECIFIC Targets on specific population SUBJECTIVE Does not target on a specific population Results are biased due to cultural Results are not biased due to cultural influence influence Population which is influenced by Administered to identify innate cultural elements display either low or abilities that is not affected by culture high scores in association with test norms PsyQuesta Learn Psychology with Afa VI. NORM REFERENCED AND CRITERION REFERENCED NORM REFERENCED Measures performance of one group of the test takers against another group of test takers. Each skill is tested by less than four items. The items vary in difficulty. CRITERION REFERENCED Measures performance of test takers against the criteria covered in the curriculum. Each skill is tested by less than four items to obtain an adequate sample of the students. Norm- Referenced test scores are Criterion - Reference tests need not be reported in a percentile rank. administered in a standardized format. If a test taker ranks 95%, it implies The score determines how much of the that he/she has performed better than curriculum is understood by the test 95% of the other test takers taker PsyQuesta Learn Psychology with Afa MODULE- 3 TEST CONSTRUCTION AND ADMINISTRATION A. GENERAL STEPS TO TEST CONSTRUCTION I. PLANNING The first step in construction of a test is careful planning. At this stage the test constructor specifies the broad and specific objectives of the test in clear terms decides upon the nature of the content or item to be included, the type of instructions to be included, the method of sampling, a detailed arrangement of preliminary administration and the final administration, a probable length and time limit for the completion of the test, probable statistical method to be used, etc. Planning also includes the total number of reproduction of the test to be made and a preparation of manual. II. WRITING DOWN THE ITEMS The second step in the test construction is the preparation of the items of the test. According to Bean, an item is defined as the single question or task that is not even broken down into any smaller units. Item writing starts with the planning done earlier the test constructor decides to prepare an essay test, the essay tests are written down. However, if he decides to create an objective test, he writes down the objective items such as the alternative response items, matching item, multiple choice items, completion item, pictorial form of items, etc. Depending on the purpose he decides to write any of their objective types of items. There are some essential prerequisites which must be made if the item writer wants to write good and appropriate items. These requirements are enumerated as follows The item writer must have a thorough knowledge and complete mastery of the subject matter. The item writer must be fully aware of those person for whom the test is meant. He must be aware of intelligence level of those persons so that he may manipulate the difficulty level of the items for proper adjustment with their ability level. The item writer must be familiar with different type of items along with their advantages and disadvantages. He must also be aware of the characteristics of good items and the common probable errors in writing items. The item writer must have a large vocabulary. He must know the different meanings of the word so that the confusion in writing the items may be avoided. He must be able to convey the meaning of items in the PsyQuesta Learn Psychology with Afa simple possible language. After writing down the items, they must be submitted to a group of subject experts for their criticisms and suggestions, which must then be duly modified. The item writer must also cultivate a rich source of ideas for items. The common source for this are textbooks, journals, discussions, questions for interview, course outlines and other instructional materials. After the items have been written down, they are renewed by some experts or by the item writer himself and then arrange in the order in which they have to appear in final. Generally items are arranged in an increasing order of difficulty and those having the same form (i.e, alternative form matched) and dealing with the same contents are placed together III. ITEM ANALYSIS MEANING AD PURPOSE After item writing, revision and editing they are subjected to a procedure called item analysis Item analysis is a technique through which those items which are valid and suited to the purpose are selected and the rest are either eliminated or modified to suit the purpose It demonstrates how effectively a given test item functions within the total test. The validity of the whole test is depend upon the validity of the individual item. It is a set of procedures that provide us with the estimate of validity of each item. The main objective of item analysis are : Item analysis indicate which items are difficult, easy, moderately difficult, or moderately easy. In other words it provides an index of the difficulty value to each item. It also provides indices of ability of the item to discriminate between inferior and superior. In other words item analysis indicates the discrimination value of each item, this is known as item validity. It indicates the effectiveness of the distractors in multiple choice items since multiple choice items are the more powerful and flexible objective items and many of standardized tests unlike this form of item. Although item analysis is done to indicate the extend to which the distractors or foils are effective in each item. It sometimes also indicates why a particular item in the test has not functioned effectively and how this might to modified so that its functional significance can be increased. ITEM DIFFICULTY In item analysis the 1 st step is to find out the difficulty value of the item or PsyQuesta Learn Psychology with Afa index of difficulty of an item. The difficulty value of the item is defined as the proportion or percentage of examines or individuals who answer the item correctly. this propotion or percentage is known as the index of difficulty of an item. If an item was answers correctly by 90% of the examines it obviously means that the item is relatively easy and it is not well discriminating. Maximum discrimination is possible when an item is answered correctly by 50% and wrongly by 50%. The proportion passing an item is inversely related to the difficulty of an item. The higher the proportion or percentage getting the item right (i.e, higher the index of difficulty) the easier the item and lower the proportion getting the item right (i.e, lower the index of difficulty) the more difficult the item. Any test to be called a measuring instrument must have some items of higher indices of difficulty, the items must have a normal distribution with respect to these indices of difficulty. The reason why items of moderate difficulty indices are preferred is that they signal the maximum variance. As the index of the item increases / decreases the variance of the item gradually decreases. i.e, its ability to make comparisons among those those who pass and those who foil decreases. Basically these are 2 important methods of determining the difficulty value of an item.: i. Method Of Judgement In this method the difficulty value of an item is determined on the basis of the judgement of experts. Items are given to a group of experts with the instruction to rank the items in their increasing order of difficulty. Subsequently, the test constructor takes a final decision keeping in view the commonality of ranks assigned to each item by the different judges as the experts. The demerit of this procedure is that they cannot be fully reliable and objective. ii. Emperical Method Also called the statistical method It is the basic and scientific method of determining difficulty index of an item. There are 2 common statistical method through which difficulty index can be estimated. One method is on the basis of response of all the examines whereas another method is on the basis of the response of only a position of the examines. P = R/N P - difficulty index R - no: of examine who pass the item. N - total no: of examines PsyQuesta Learn Psychology with Afa I. ITEM DISCRIMINATION The determination of the index of discrimination is also known as validity index. Index of discrimination is that ability of the item on the basis of which the discrimination or distinction is made between superiors and inferiors. Bean has defined this index as “the degree to which the single item separates the superior from the inferior individual in the trait or group of traits being measured.” From the point of view of the discriminatory power, all test items can be divided into items that are either (a) positively discriminating, (b) negatively discriminating or (c) non discriminating A positively discriminating item may be defined as one in which proportion or percentage of correct answer is higher in the upper group. When the proportion or percentage of correct answer is lower in the upper group is called the negatively discriminating item. Items in which the percentage or proportion of correct answers is approximately equal in both the groups is called a non discriminating item. The negative and non discriminating items are dropped after item analysis. IV. PRELIMINARY ADMINISTRATION When the items have been written down and modified in the light of the suggestions and criticism given by the experts, the test is said to be ready for its experimental try out. The purpose of preliminary administration of any psychological test is as given below: i. Finding out the major weaknesses, omission, ambiguities and inadequacies of the items. ii. Determining the difficulty values of each item which in turn helps in selecting items for their even and proper distribution in the final form. iii. Determining the validity of the each individual item. The preliminary administration helps in determining the discriminatory power of each individual item. iv. Determining a reasonable time limit of the test. v. Determining the appropriate length of the test, i.e, it helps in determining the no: of items to be included in the final form. vi. Determining the inter-correlations of item so that overlapping can be avoided. vii. Identifying any weakness and vagueness in directions or instructions of the test as well as in the fore-exercises or sample questions of the test. According to Conard (1951), there should be at least three preliminary administration of the test 1. 1st administration to detect any gross defects. Ambiguities and omission in items and instructions. For this, the no: of examines should be PsyQuesta Learn Psychology with Afa not less than 100. 2. 2nd administration to provide data for item analysis, and for this the no: of examines should be around 400. the sample for this must be similar to those for whom the test is intended. It aim at obtaining three kinds of information regarding the items : i. The difficulty value of the item ii. The discrimination index of the item iii. The effectiveness of distractors. 3. The 3rd preliminary administration is carried out to detect any minor defects that may not have been detected by the first two preliminary administration. At this stage the items are selected after item analysis and they constitute the test in the final form. The 3rd administration indicates how effective the test will really be when it would be administered on the sample for which it is really intended. V. STANDARDIZATION Norms are established for the sake of standardization of any test. Standardization is the process whereby a test is administered to a representative sample of population whom the test is meant for, for the sake of establishing norms. A standardized test is the one that has normative data, as well as clearly specified administration and scoring procedures. VI. PREPARATION OF MANUAL AND REPRODUCTION OF THE TEST The last step in test construction is the preparation of a manual of the test. In the manual the test constructor reports the psychometric properties of the test, norms and references. Thus gives a clear indication regarding the procedures of the test administration, the scoring methods and time limits if any, of the test. It also includes instructions as well as the details of arrangement of materials i.e, whether the items have been arranged in random order or in any other order. In general, the test manual should yield information about the standardisation sample, reliability, validity, scoring as well as practical considerations The test constructor after seeing the importance and requirement of the test finally orders for printing of the test and the manual B. RELIABILITY AND ITS TYPE Reliability is one of the important characteristics of any test. it refers to the precision or accuracy, of the measurement of scores in other words the score should be consistent. When all other factors are held constant or somehow controlled, a reliable test is one that produces identical results for an examinee from one occasion to the other. PsyQuesta Learn Psychology with Afa Test-Retest Reliability Test-retest reliability deals with two performances of the same test by the same persons on two different occasions with a reasonable time gap. This way of two administrations of the same test yield two independent set of scores. The two sets when correlated, give the value of the reliability coefficient. The test-retest coefficient is also known as the coefficient of stability. The test takers' scores on first administration of the test are correlated with their scores obtained on the second administration of the same test. a major advantage of this method is that there is maximum control over the test taker and test item variables. Alternate-Form Reliability Also known as parallel-form reliability, equivalent form reliability, and the comparable forms reliability. In order to overcome the problems caused by using the same version of the test on two occasions. This involves computing the alternate-form reliability. In this approach the test developer develops two alternate or parallel forms of the same test. Ideally speaking the two forms should be independently developed and completely parallel or equivalent forms of the same measure. They should match each other in all respects. Two forms of the test are administered to the same sample, either immediately the same day or with the time interval of usually a fortnight. If reliability is calculated on the basis of data collected immediately on the basis of two administrations of the test, it is called alternate form (immediate)reliability and if reliability is calculated on the basis of data collected after a gap of fortnight, it is called alternate form (delayed) reliability. Internal consistency reliability Also called as split half reliability. Indicates the homogeneity of the test Alternate form reliability is a popular type of reliability but with some obvious limitation like it is not easy and requires a lot of time and effort, the practice effect and prior experience may affect performance on the second occasion. The time gap between two administrations is another intervening variable. To overcome these problems, split-half reliability is computed. The test is arbitrarily divided into two halves. Scores on one half are correlated with scores on the other, using Pearson's r. The obtained value is the 'coefficient of internal consistency. In this approach error caused by the temporal variation or different sets of items is controlled. The common method of splitting the test is odd- even method. PsyQuesta Learn Psychology with Afa Inter-scorer Reliability/Inter-rater Reliability At times the test may have to be evaluated by more than just one examiner. In case of objective tests, a test taker will obtain the same marks no matter who, and how many people, evaluate it. On the other hand, when essay type or other open-ended items are to be evaluated, a need for multiple scorers may arise. When more than one person is evaluating, inter-scorer reliability is required. Some common procedures include the following: 1. Two examiners evaluate and mark a test of a group of test takers. The resulting two sets of scores are used for computing correlation between the two sets Intra-class coefficient/ coefficient of concordance. Two or more examiners score the performance of a number of test takers and the coefficient is computed from these scores. C. MEANING OF VALIDITY AND ITS TYPES The term validity means truth or fidelity. Thus, validity refers to the degree to which a test measures what it claims to measure. a test is valid to the extent that it serves the purpose for which it is to be used. Types Of Validity Content Validity: Content validity is also known as curricular validity intrinsic validity, relevance, circular validity and representativeness. Content validity is a non-statistical type of validity that is usually associated with the achievement test. When a test is constructed so that its content of term measures what the whole test claims to measure, then the test is said to have content validity. For example, suppose a professor wants to test the overall knowledge of his students in the subject of elementary statistics, test will have content validity if the test covers every topic of elementary statistics that he taught in the class. Face validity Face validity often confused with content validity Face validity refers not what the test actually claims to measure but to what it appears to measure superficially. in other words face validity is the mere appearance that the test has validity Thus, face validity should not be taken in the technical sense, nor should it be regarded as a substitute for objectively determined validity, when a test item looks valid to the group of examinees because it provides a logical link with the objective, the test is said to have face validity. PsyQuesta Learn Psychology with Afa Criterion-Related Validity One approach to assessment of validity is through comparing the test results with a criterion. Consider a situation where you are looking for a foot ruler to measure the length of a large sheet of paper. Someone brings a stick to you and tells you that the stick is foot long and you can use it instead of a ruler. You doubt this and want to ensure that this stick is really foot long. Now you look for another object that you are sure is foot long, though cannot be used for measuring another object. You see that the tiles in your kitchen are 6x6 inches square. You measure the stick against the tiles, and you see that the stick is equal to two tiles in length i e.. 12 inch long. What you have done here is, you measured your tool with reference to a criterion and the result of this exercise showed that your tool was valid. You can use it instead of a ruler. There are two types of criterion-related validity: Predictive Validity: "The evidence that a test forecast score on the criterion at some future time", Predictive validity is also known as empirical validity or statistical validity. In predictive validity, a test is correlated against the Criterion to be made available.sometime in the future. The test scores are obtained and then a time gap of month or years is allowed to elapse, after which the Criterion scores are obtained. Concurrent validity: Concurrent valid evidence is defined as "evidence for criterion validity in which the test and the criterion are administered at the same point in time Construct Validity Construct validity is measured keeping in view the particular construct that the test is supposed to measure. If the test is found to measure the construct in question, then it is a valid test, and it does not measure the trait then the test is not accepted as valid. Construct validity is about ensuring that the method of measurement matches the construct you want to measure. D. NORMS AND ITS TYPES A norm represents a typical level of performance for a particular group. In a test, norm is that score which has been obtained most by the group. Types Of Norms Percentile Norm Percentile norm indicates, for each raw score, percentage of standardization sample that falls at or below that raw score. PsyQuesta Learn Psychology with Afa For example, suppose Mohan has a score of 26 the mechanical reasoning test and if 40% of the standardization sample secures below the score of 26 Mohan has percentile rank of 40 or PR40 and percentile score of 26. Percentile norms, provide a basis for interpreting an individual's score on a test in terms of his own standing in a particular standardization sample. If the percentile norm is to be meaningful, it should be based upon a sample which has been made homogenous with respect to age, grade, sex occupation and other factors. Standard score norm Norm based upon a standard score here unit of the scale are equal so that they convey the same meaning throughout the whole range of the scale. Standard score, like the percentile score, is a derived score, it has a specified or fixed mean and fixed standard deviation. Different types of standard scores are z-score, t-score, stanine score etc. Standard scores are needed primarily for two reasons. Firstly, when the performance of the same person on different tests to be compared, and it is best done through converting the raw scores into standard scores. Secondly, standard scores have equal units of measurement, and their size does not vary from distribution to distribution. Hence, they are frequently used in interpreting test scores. Age Equivalent Norms Age equivalent norms are defined as the average performance of a representative sample of a certain age level on the measure of a certain trait or ability. For example, if we measure the weight of a representative sample of 10year-old girls of the state of Bihar and find out the average of the obtained weight we can determine the age norm for the weight of 10-year-old girls. Age norms are more suited to those traits or abilities which increase systematically with age. So, most of the physical treats like weight, height etc and cognitive abilities like intelligence shows such systematic change and age norms can be more appropriately used for these traits Grade-Equivalent Norms grade equivalent norms are defined as the average performance of a representative sample of a certain Grade or class. The test whose norms are being prepared, is given to the representative sample selected from each of the several grades or classes, After that the average performance of each grade on the given test is determined. Grade equivalents represent the scores on educational achievement tests attained by children in a certain grade. These norms are obtained by calculating the mean raw scores of children in the standardization sample representing each grade. If 6th grade children in the standardization sample PsyQuesta Learn Psychology with Afa obtained a mean score of 35 in arithmetic test, then this raw score has a grade equivalent of 6. Hence a student obtaining 35 on the same test will be said to have a grade equivalent of 6. t SCORE: -Standard score with the mean of 50 and standard deviation of 10 PsyQuesta Learn Psychology with Afa MODULE - 4 BASICS OF PSYCHOLOGICAL RESEARCH A. MEANING AND CHARACTERISTICS OF SCIENTIFIC RESEARCH Scientific research is a systematic and objective attempt to provide answers to certain questions. scientific research may be defined as the systematic and empirical recording and analysis of controlled observation, which may lead to development of theories, concepts, generalizations and principles, resulting in prediction and control of these activities that may have some cause effect relationship. Kerlinger(1973) defined scientific research as “a systematic, controlled, empirical and critical investigation of hypothetical propositions about the presumed relations among natural phenomena.” B. CHARACTERISTICS OF SCIENTIFIC RESEARCH: 1. directed towards the solution of a problem or any relation between variables under study. 2. 3. Always based upon empirical or observable evidences. 4. Involves precise observation using valid and reliable instruments for collection of data and uses statistical methods for accurate description of the results obtained. 5. Emphasizes on development of theories, principles and generalizations regarding the variables studied and the populations. 6. systematic, objective and logical procedures. 7. Research is marked by patience, courage and unhurried activities. 8. Requires full expertise and awareness of relevant aspects of the problem, history, literature and sophisticated statistical methods for analysis. 9. The designs, procedures and results of scientific research must be replicable to ensure reliability and validity. 10. The research requires skills of writing and preparing a research report. C. TYPES OF RESEARCH 1. Historical research: investigates, records, analyses and interprets the events of the past for the purpose of discovering sound generalizations that are helpful and useful in understanding the past, the present and to a limited extent, the anticipated future. 2. Descriptive/non-experimental/Correlational research: describes, records, analyses and interprets the existing conditions. Here, an attempt is made to discover relationships that exist between non-manipulated variables, or some comparison or contrast among them. PsyQuesta Learn Psychology with Afa 3. Experimental research: one in which primary focus is upon variable relationship. Here, certain variables are controlled or manipulated and their effect is examined upon some other variables. 1. HISTORICAL application of scientific method to the description and analysis of past events. It is supposed to be a factual integrated account of the relationships between persons, times and events. Historical research becomes necessary not only for knowing the past but also for understanding the present and predicting future in that context. Steps in historical research: 1. Selection of the problem – all problems are not suitable for historical research. Only those problems that can be studied with the help of historical records and have some social utility can be picked up. 2. Formulation of hypothesis – they are explicitly formulated in historical research. Researcher gathers evidence and carefully evaluates the trustworthiness of the hypothesis. 3. Collection of data – historical data is collected from primary sources and secondary sources. Primary sources are eyewitness accounts by actual observers or participants in the event. Secondary sources are gathered after the event talking to the observers and participants. 4. Analysis and generalization – historians employ principles of probability and rigorously subject the evidence to critical analysis. 2. DESCRIPTIVE RESEARCH V/S EXPERIMENTAL RESEARCH descriptive research is where independent variables cannot be manipulated and cannot be experimentally studied. The subjects are not randomly assigned into different treatment conditions. the response of a group of subjects are measured on one variable and then compared with their measured responses on another variable. Such research is also called R/R research because changes in one set of responses are compared with possible changes in another set of responses. Field studies, ex post-facto research, survey research, content analysis, case studies etc are examples. An experimental research is where the independent variables can be directly manipulated by the experimenter, and participants or subjects are randomly assigned into different treatment conditions. experimental research is sometimes called the S/R research because the researcher manipulates a stimulus in order to establish whether or not this produces a change in certain response. They are further of two types – laboratory experiments and field experiments. PsyQuesta Learn Psychology with Afa 3. BASIC VERSUS APPLIED RESEARCH Basic or fundamental or pure research is the formal and systematic process where the researcher’s aim is to develop a theory or a model by identifying all the important variables in a situation. They try discovering broad generalizations and principles about those variables. It utilizes a careful sample so that its conclusion can be generalized to immediate situations but has little concern with the actual application of its generalizations or principles. Applied research, as its name implies, applies the theory or model developed through the basic research to the actual solution of problems. It has characteristics common to fundamental research. It also tends to make generalizations and uses various sampling methods. The main purpose of applied research is not to develop theories about a fact but to test those theories in actual situations. D. RESEARCH PROCESS Even though research is a tedious, painful and slow moving job, following certain steps in conducting the research the work can be carried smoothly and with least difficulty. 1. IDENTIFYING THE PROBLEM The first step in conducting a research is to identify the problem. The researcher must discover a suitable problem and define it operationally. A problem is defined as that interrogative testable statement which shows a relationship between two or more variables in an unambiguous manner. For example, take the following statement:What is the relationship between academic ability and socio- economic status? Characteristics of a problem statement: 1. A problem statement is clear, specific, unambiguous and substantially relevant 2. It expresses the relationship between two or more variables 3. A problem statement is testable with empirical methods 4. It must avoid moral or ethical judgments that are difficult to study 5. It can be general or specific with sufficient importance Ways in which problems manifest: 1. When there is a clear gap in the results of several investigations in the same field 2. When results of several studies disagree with each other 3. When the facts of a field are based on unexplained information Types of problems Research problems can be broadly categorized into two types: PsyQuesta Learn Psychology with Afa a. Solvable problem A solvable problem raises questions which can be answered with the use of our normal capacities. It has two features; It can be solved by empirical methods that use observation of natural events It is possible to advance a suitable hypothesis as a tentative solution to it b. Unsolvable problem It is one that raises questions which cannot be answered using our normal capacities. They usually pertain to supernatural phenomena. It has three characteristics as follows; An unsolvable problem possess the trait of un-structuredness as the researcher’s interest is not clear but unorganized It has inadequately defined terms and operational definition and variable contains vagueness It is difficult to collect relevant data regarding the problem 2. HYPOTHESIS When the problem has been stated, a tentative solution in the form of a testable proposition is offered by the researcher which is called a hypothesis. According to McGuigan (1990) hypothesis is “a testable statement, of potential relationship between two (or more) variables, that is advanced as potential solution to the problem.” On the basis of given definition, two points can be suggested about hypo thesis. First, a hypothesis is a testable statement where measurable variables are used. Second, a hypothesis exhibits either a general or specific relationship between variables. After testing the hypothesis the researcher may find it to be correct or incorrect. A good hypothesis meets the following criteria or characteristics; 1. The hypothesis should be conceptually clear 2. The hypothesis should be testable 3. The hypothesis should be economical and parsimonious 4. The hypothesis must be related to the existing body of theory and facts 5. The hypothesis should have logical unity and comprehensiveness 6. The hypothesis should be general in scope 7. The hypothesis should be related to available scientific tools and techniques 8. The hypothesis should be in accord with other hypotheses in the field Types of hypothesis Alternative Hypothesis (research or experimental hypothesis) - The alternative hypothesis states that there is a relationship between the two variables being studied (one variable has an effect on the other). It states that the results are not due to chance and that they are significant in terms of supporting the theory being investigated. PsyQuesta Learn Psychology with Afa Null Hypothesis - The null hypothesis states that there is no relationship between the two variables being studied (one variable does not affect the other). It states results are due to chance and are not significant in terms of supporting the idea being investigated. Directional Hypothesis - A one-tailed directional hypothesis predicts the nature of the effect of the independent variable on the dependent variable. E.g., adults will correctly recall more words than children. Non-directional Hypothesis - A two-tailed non-directional hypothesis predicts that the independent variable will have an effect on the dependent variable, but the direction of the effect is not specified. E.g., there will be a difference in how many numbers are correctly recalled by children and adults. Simple hypothesis: contains only one or two variables Complex hypothesis: hypothesis that contains more than two variables and require complex statistical calculations 3. VARIABLES variables may be defined as those attributes of objects, events, things and beings which can be measured. variables are the characteristics or conditions that are manipulated, controlled or observed by the experimenter. Intelligence, anxiety, aptitude, income, education, authoritarianism etc are examples. Variables can be classified in several ways. Some of the commonly accepted classifications are given below. a) Dependent variables and Independent variables The dependent variable is defined as one which experimenter makes a prediction. The independent variable is defined as the one which is manipulated, measured and selected by the researcher to produce changes in the measured behaviour (dependent variable). The dependent variable is the characteristic or condition that changes as the experimenter manipulates the independent variables. The independent variable is that characteristic or condition which is manipulated or selected by the experimenter to find its relationship in some observable phenomena. For example if a researcher wants to find the effect of teaching methods on academic achievement, teaching methods constitute the independent variable and the academic achievement constitute the dependent variable. Extraneous variables: are the uncontrolled variable that may affect the dependent variable. The experimenter is not interested in finding the extraneous variable’s effect and thus has to try and control them as far as possible. They are also referred to as confounding variables or relevant variables. 4. FORMULATING RESEARCH DESIGN it is the blueprint of the detailed procedures of testing the hypothesis and analyzing the obtained data. PsyQuesta Learn Psychology with Afa the sequence of those steps taken ahead of time to ensure that the relevant data will be collected in a way that permits objective analysis of the different hypotheses. The selection of any research design is based upon the purpose of the investigation, types of variables and the conditions in which the research has to take place. Basically a research design serves two functions. i. it answers the research question as objectively, validly and economically as possible. ii. It acts as a control mechanism. This enables the researcher to control unwanted variances such as experimental variance, extraneous variance and error variance. Types of research design I. Within-groups design (repeated treatment/measures) If the researcher is using only one group of subjects who will be tested under different values of the independent variables, the resulting experimental design is called within-groups design. Such designs are of two sub types. a. Complete within-groups design - in which practice effects are balanced by administering the conditions several times to each subject, in different orders each time, to obtain interpretable results. b. Incomplete within-groups design – in which each condition is administered to each subject only once while varying the order of administration across the subjects in a way the practice effects can be neutralized. II. Between groups design (between subjects/independent measures) If the researcher decides to use two separate groups for each value of the independent variable, the resulting design is referred to as between-groups design. a. Randomized group design – one in which subjects are randomly assigned to the different groups meant for the different conditions or values of independent variable. b. Matched groups design – one in which the subjects are matched depending upon mean, standard deviation, pairs etc. c. Factorial design – in which two or more independent variables are studied in all possible combinations (each having a separate group) in order to study their independent and interactive effects on dependent variable. 5. REVIEWING LITERATURE A collective body of works done by earlier scientists is technically called the literature. Any scientific investigation starts with a review of the literature. The main objectives of literature review are given below. PsyQuesta Learn Psychology with Afa 1. Identifying variables relevant for research: when the researcher makes a careful review of the literature, he becomes aware of the important and unimportant variables in the concerned area of research. 2. Avoidance of repetition: helps the researcher in avoiding any duplication of work done earlier by someone. Prior studies have to be the foundation of the current study and in some cases replication of a study also can be needed. 3. Synthesis of prior works: enables the researcher to collect and synthesize the findings of previous studies related to the problem. It gives a wider perspective to the future research. 4. Determining meaning and relationship among variables: careful review of literature enables the researcher in discovering significant variables and identifying important relations among variables. Sources of review Journals and Books- Different research journals and books relevant to the areas of interest are the primary sources of literature review. Libraries have books and periodical sections with journals. Referred journals carry only those articles which are carefully reviewed by experts before publication. Reviews- Reviews are short articles that give brief information regarding the work done in a particular area over a period of time. Reviewers select research articles of their interest, organize them content-wise, criticize their findings and offer their own suggestions and conclusions. They can often guide you to serious books and journals. Abstracts- In an abstract, researchers can get all the relevant information regarding the paper and the problem discussed. Only limitation of abstracts is that the researcher who desire detailed information regarding the methodology and results. Internet- Today, the internet is a very easy and quick source of literature review. They provide an easy way to find the original articles related to the problem. Plenty sites that discuss issues relevant to the problem can give enough information and resources. Doctoral dissertations- Dissertations by those who studied similar issues can be found at many institutes and libraries. The researcher can choose interested topics and find useful information and references there. Most dissertations include chapters like an introduction, review of the literature, purpose of the study, method of the study, data collected, analysis and interpretation, summary and conclusion. Supervisors /research professors- Those who have been working with researchers for long time often know the literature well and are able to guide in right direction. Therefore, it is advised that a researcher consult supervisor, research professors and assistants in the process. 6. SAMPLING PsyQuesta Learn Psychology with Afa A population may be defined as the identifiable and well specified group of individuals or objects the researcher wants to study. A population can be finite or infinite like the students of a class and fishes in a river respectively. A measure based upon the entire population is called a parameter. A sample is any number of units selected to represent the population according to some rule or plan. A measure based upon a sample is known as statistic. most sampling methods can be categorized into two – 1. Probability Sampling Methods specify the probability or likelihood of inclusion of each element or individual in the sample. That means i. the size of parent population from which the sample has to be taken is known to the researcher. ii. each element of the population must have an equal chance of being included in a subsequent sample. iii. the desired sample must be clearly specified. The major probability sampling methods are the following: a. Simple random sampling- every member of a population has an equal chance of being a sample i.e., samples are randomly assigned b. Systematic sampling- every ith item from the list is selected as the sample. c. Stratified random sampling - used in cases where the data is heterogeneous. The population is divided into different strata with specific characteristics d. Area or cluster sampling- the total population is divided into different clusters\ groups and sample units are randomly from each of these clusters 2. Non-probability Sampling Methods there is no way of assessing the probability of the elements of the population being included in the sample. Important techniques of non-probability sampling are: a. Quota sampling- the sample size is fixed first and then quota is fixed for various categories of population b. Accidental sampling- investigator selects samples based upon his convenience, economy and not based on specific traits c. Judgemental/purposive sampling- Sampling units are selected on the basis of some belief of a researcher that it represents the population well. d. Snowball sampling-Initially selected respondent provides addresses of additional respondents. e. Saturation sampling - Drawing all elements of individuals having characteristics of interest to the invigilator is called saturation sampling PsyQuesta Learn Psychology with Afa 1. 2. 3. 4. 5. 6. FUNDAMENTALS OF SAMPLING Universe or population – population is a group of objects, animate or inanimate, finite or infinite under the study Sampling frame – the list of all items in a population Sampling design – it means a plan for obtaining a sample from the sampling frame. It refers to the procedure the researcher would adopt in selecting the sample. Statistic and parameter –statistic is a numerical value based upon the sample ,where as parameter is the numerical value based upon population Sampling error – Sampling error is a difference between the parameter and statistic, ie “a sampling error is a difference between parameter and estimate of that parameter which is derived from sample” Confidence level and significance level – confidence level is the likelihood that the actual value will fall within the stated precision limits. Significance level, on the other hand, is the level that indicates the likelihood that the answer will fall outside the precision range. If the confidence level is 95%, then the corresponding significance level is 100-95 = 5% or 0.05. E. DATA COLLECTION TECHNIQUES 1. QUESTIONNAIRE AND SCHEDULE Questionnaires are used when factual information from the respondents is desired. It consists of a form containing a series of questions where the respondents themselves fill in the answers. It enables researchers to get first-hand information regarding the vagueness of items and establish a warm relationship with the persons being tested. A schedule consists of a form containing a series of questions, which are asked and filled in by the investigator in a face to face situation. An opinionnaire is an information form which attempts to measure the attitude or belief of an individual. Hence, an opinionnaire is also called a attitude scale. Types of Questionnaires 1. Fixed response questionnaire/ close ended consists of statements or questions with a fixed number of options or choices. Example: Do you feel shy talking to members of opposite sex? (Yes/No) Data collection and statistical analysis will be made easier with this type of questionnaires. They will also be time- money and energy saving. Disadvantages include the researcher’s inability to provide enough relevant responses and these questionnaires can encourage a respondent to adopt some kind of response set or bias. 2. Open end questionnaire questions that require short or lengthy answers by the respondents. particularly useful in situations where researchers have little information PsyQuesta Learn Psychology with Afa about the subjects to be studied elicits unanticipated and insightful responses that can increase the researchers’ understanding. Disadvantages include biases that may occur. Educational or socio- economic backgrounds may affect the way different respondents understand and respond to items. They are also time consuming and may reduce participation, scoring and statistical analysis are more difficult and misleading. Example: What are the causes of student unrest in the campus? Characteristics of a good questionnaire: 1. Questionnaire must be concerned with specific and relevant topics. 2. Significance, objectives and aims of the questionnaire are clearly stated 3. Be short and avoid burdening the respondent 4. Use simple and clear language for directions and questions 5. Questions must be objective without hinting any possible answer 6. Embarrassing questions, presuming questions and hypothetical questions must be avoided 7. The items must be placed in a good and systematic order to elicit valid responses 8. Must be clearly arranged, printed attractive and neatly 2.INTERVIEW Interview is a face to face interaction between interviewer and interviewee which intends to elicit some desired information from the latter. Like questionnaire, interview aims to obtain data regarding the respondents with minimum bias and maximum efficiency. Success of an interview depends on characteristics of the interviewer, the questions and the respondent. Types of interviews 1. Structured or formal interview – already prepared questions are asked in a set order by the interviewer and answers are recorded in a standardized form. since questions and response marking are structured, a relatively less trained interviewer can conduct the interview smoothly. The systematic procedure also makes them more valid than informal interviews. At the same time they are time consuming and expensive, also less valid than other methods like bio data analysis and standardized psychological test. 2. Unstructured or informal interview – there are no pre- determined questions nor any preset order of the questions and it is left to the interviewer to ask some questions in a way he likes regarding a number of points the interview is supposed to build up on. An advantage to informal interview is that the interviewer gets to probe more to gain a deeper understanding of the respondent’s behaviour. Disadvantages include the dependency on the interviewer that reduces reliability, need for skilled and trained interviewer and the data collected being difficult to quantify and analyse statistically. PsyQuesta Learn Psychology with Afa 3. 1. 2. 3. 4. 5. Advantages of Interviews Allows greater flexibility in the process of questioning Facilitates obtaining desired information readily and quickly Ensures the interviewee interpret and answer the questions themselves Validity of the verbal responses checked non-verbally Desired level of control can be exercised over the situations and contexts Disadvantages of Interviews 1. Interviewers’ variability 2. Inter-interviewer variability 3. Doubtful validity and dependability of verbal responses 4. Time consuming 5. Variations inherent to the interviewing context 6. No foolproof system for recording responses Important sources of error in interviews 1. Attitude of the interviewer 2. Incomprehensibility of the questions asked 3. Lack in warmth in the interview situation 4. Lack of motivation of respondents 5. Duration of the interview 3. CONTENT ANALYSIS Content analysis\ document analysis, is a method used to analyse qualitative data. it is a technique that allows a researcher to take qualitative data and to transform it into quantitative data (numerical data). the analyzer takes the communications or documents prepared by the respondents/subjects and systematically finds out the frequency or proportion of their appearances. The technique can be used for data in many different formats, for example, interview transcripts, letters, diaries, journals, stories, reports, academic writings, autobiographies, film, and audio recordings. It can also be used for analyzing responses to Projective tests. Methods of content analysis The most widely used method for content analysis was proposed by Berelson (1954) and has three steps. 1. Specification of the universe – what are we looking for (variables) 2. Identify units of the analysis – five major units are words, themes, items, characters and space-and-time measures 3. Quantification – assigning numerals to the objects of analysis 4. OBSERVATION refers to watching and listening to the behaviour of other persons over time without manipulating or controlling it and record findings in a way that allow some degree of analytical interpretation and discussion. PsyQuesta Learn Psychology with Afa Observation involves broadly selecting, recording and analysing behaviour for empirical aims of description or development of theory. Observation usually occurs in natural settings although it could be in contrived settings as laboratory experiments and simulations. It captures significant events or occurrences affecting relations among the observed. It identifies regularities and recurrences in social life through comparing and contrasting other findings and theories. 1. 2. 3. 4. 5. 6. 7. Types of observation Systematic observation – done according to some explicit procedures as well as in accordance with the logic of scientific inference. Unsystematic observation – casual observation made by the investigator without specifying any explicit and objective inference. Participant observation – the investigator actively participates in the activities of the group to be observed. Also called disguised participant observation. Observations occurs in a natural setting enabling to record the behaviour in a realistic manner. observations of longer duration can yield broad and meaningful information for understanding human behaviour. It is usually unstructured and fails to be precise about data collection, making statistical analysis difficult. It is time consuming and makes the process effortful to observers. Since the participation is active, the observer sometimes develop emotions towards the observed making the data collected biased. Non-participant observation – the investigator observes the behaviour of subjects in their natural setting without any participation. usually structured, the observer plans for the nature of the setting they want to observe, representativeness of the data, and problems associated with the presence of observer. the obtained data is more reliable and representative. The observer can plan the aspects and process of observation in a nice way. The observer is also able to concentrate upon specific aspects of social behaviour and gets better chances to find data or solution regarding the probe.Limitations of this type of observation are (1) that the observed group and their setting does not remain a natural one with the observer around and (2) it captures less natural context of social settings than participant observation. Naturalistic observation - This technique involves observing involves studying the spontaneous behaviour of participants in natural surroundings. The researcher simply records what they see in whatever way they can. Controlled observation - The researcher decides where the observation will take place, at what time, with which participants, in what circumstances and uses a standardized procedure. Participants are randomly allocated to each independent variable group. Subjective observation - In subjective observation, which is also called introspection, a person observes his own mental activities. Introspection literally means, "Looking within". It means getting insight into one's own mental activities. In introspection, a person perceives, analyses, and gives a PsyQuesta Learn Psychology with Afa 8. report of his own feelings and experiences. 5. RATING SCALE Rating scale is a technique to assess both actual behaviour as well as remembered behaviour. It is defined as a technique through which the observer or rater categorizes the objects, events or persons on a continuum, represented by as series of continuous numerals. The purpose of rating scale is to know what kind of impressions the objects or persons have made upon the raters. A rating scale usually has two, three, five, seven, nine or eleven points on a line with descriptive categories. rating scales are divided into few categories. 1. Numerical scale - the observer is supplied with a sequence of numbers well defined as to what they mean. 2. Graphic scale - the scales are presented graphically in which descriptive cues corresponding to the different scales given. 3. Percentage rating - requires the rater to place the rates among different specified percentage groups. 4. Standard scale is one in which the rater is presented with some standards with pre-established scale values. 5. Scale of cumulated points - based on cumulated or summated points. Here the person’s total score is the sum of individual ratings assigned to all items on the scale. 6. Forced choice rating scale - the rater is given a set of attributes in terms of verbal statements for a single item and s/he gets to decide which one represents the rated object appropriately. F.CARRYING OUT STATISTICAL ANALYSIS It is customary to choose the appropriate statistical tests on the basis of the obtained data. If the data fulfil the requirement of parametric assumptions, any of the parametric tests which suit the purpose can be selected. If the data do not fulfil the parametric requirements, any of the non-parametric tests could be used. Other things to be kept in mind in selection of appropriate statistical tests are the number of independent and dependent variables and the nature of the variables, that is, whether they are nominal, ordinal, ratio or interval. When both independent and dependent variables are interval measures and are more than one, multiple correlation is the most appropriate statistic. On the other hand, when they are interval measures and their number is one, Pearson r may be used. With nominal and ordinal measures the non-parametric statistics are the common choice. Sometimes, researchers transform the measures so that appropriate statistical test may be applied without loss of much information. For PsyQuesta Learn Psychology with Afa example, if scores of two groups on interval measures are available but the data do not fulfil the requirement for the t test, the researcher can transform the interval measures to ordinal measures and subsequently, apply the Mann-Whitney test. DESCRIPTIVE AND INFERENTIAL STATISTICS Descriptive statistics describe a sample. take a group that you’re interested in, record data about the group members, and then use summary statistics and graphs to present the group properties. With descriptive statistics, there is no uncertainty because you are describing only the people or items that you actually measure. You’re not trying to infer properties about a larger population. Inferential statistics takes data from a sample and makes inferences about the larger population from which the sample was drawn. Because the goal of inferential statistics is to draw conclusions from a sample and generalize them to a population, we need to have confidence that our sample accurately reflects the population. Statistics which enables to make inferences or generalizations about a population with known possibilities of error are called statistical inferences. any statistical analysis dealing with the entire population, or to describing the sample is descriptive statistics. On the other hand, if we deal with a sample, our analysis relates not only to the characteristics of the sample but it also provides information about the population. Such statistics are known as inferential statistics or inductive statistics or sampling statistics. DRAWING CONCLUSIONS Once a researcher has designed the study and collected the data, using statistics, researchers can summarize the data, analyze the results, and draw conclusions based on this evidence. Not only can statistical analysis support (or refute) the researcher’s hypothesis; it can also be used to determine if the findings are statistically significant. When results are said to be statistically significant, it means that it is unlikely that these results are due to chance. Based on these observations, researchers must then determine what the results mean. In some cases, an experiment will support a hypothesis, but in other cases, it will fail to support the hypothesis. Just because the findings fail to support the hypothesis does not mean that the research is not useful or informative. In fact, such research plays an important role in helping scientists develop new questions and hypotheses to explore in the future. After conclusions have been drawn, the next step is to share the results with the rest of the scientific community. This is an important part of the process because it contributes to the overall knowledge base and can help PsyQuesta Learn Psychology with Afa other scientists find new research avenues to explore. G. STRUCTURE OF A RESEARCH REPORT The writing of the research report requires imagination, creativity and resourcefulness. Research reports should be written in a dignified and objective style, although there is no one such style which is acceptable to all. The research reports aim at telling the readers the problems instigated, the methods adopted, the results found and the conclusion reached. The research paper should be written in a clear and unambiguous language so that the reader can also objectively judge the adequacy and the validity of the research. For attaining objectivity, personal pronouns such as I, you, we, my, our, etc., should be avoided and as their substitutes, expressions like 'investigator’, ‘researcher’ should be used. The research report, whether it is based on a dissertation or short-term research paper, should be of fairly standardized pattern. Such standard conventions have been neatly outlined by Publication Manual of the American Psychological Association, and have been universally accepted. The publication manual has suggestions as to how to write a report effectively, how to present ideas with concise expression and how to avoid ambiguity and increase re-adaptability of the report. The following outline or format is the typical research form prepared according to the American Psychological Association’s Publication Manual (1983); I. Title page a. Title b. Author’s name and affiliation c. Running head d. Acknowledgements II. Abstract III. Introduction (no heading) a. Statement of the problem b. Background/review of literature c. Purpose and rationale/hypothesis IV. Method a. Subjects b. Apparatus (if any) c. Design d. Procedure V. Results a. Tables and figures b. Statistical presentation PsyQuesta Learn Psychology with Afa VI. VII. VIII. Discussion a. Support or non-support of hypothesis b. Practical and theoretical implications c. Conclusions References Appendix (if appropriate) APA style of reference writing References comprise all documents including journals, books, technical reports, computer programs and unpublished works mentioned in the text of the report. References are arranged in alphabetical order by the last name of the author(s) and the year of publication in parenthesis,( ). in case of unpublished citations, only the reference is cited. Sometimes no author is listed and then, in that condition the first word of the title or sponsoring organization is used to begin the entry. When more than one name is cited within parenthesis, the references are separated by semicolons. In parenthesis, the page number is given only for direct quotations. The researcher should check carefully that all references cited in the text appear in the references and vice versa. References should not be confused with Bibliography. A bibliography contains everything that is included in the reference section plus other publications which are useful but were not cited in the text or manuscript. Bibliography is not generally included in research reports. Only references are usually included. The Publication Manual of the American Psychological Association (1983) has given some specific guidelines for writing references of various types as indicated below. 1. For reference of books with single author: Siegel, S (1956) Non-parametric Statistics for the Behavioural Sciences, New York: McGraw Hill. 2. For reference of books with multiple authors: Guilford, JP & Fruchter, B (1978) Fundamental Statistics in Psychology and Education, New York: McGraw Hill. 3. For reference of corporate or association as author: American Psychological Association (1983) Publication Manual (3rd ed.) Washington. 4. For reference of journal article: Edwards, AL & Kenny, KC (1946). A Comparison of Thurstone and Likert techniques of Attitude Scale Construction, Journal of Applied Psychology, 30, 70-73. 5. For reference of thesis or dissertation: Singh AK (1978) Construction and Standardization of a Verbal Intelligence Scale. Unpublished Doctoral Dissertation, Patna University, PsyQuesta Learn Psychology with Afa Patna. 6. For reference of chapter in an edited book: Atkinson, RC & Shiffrin, RM(1968) Human Memory: A proposed system and its control processes. In The psychology of Learning and Motivation, ed. KW Spence and JT Spence, vol. 2, pp 89-195. New York: Academic Press. PsyQuesta Learn Psychology with Afa