» In order to create a valid experiment, the experimenter can only change ONE variable at a time » The variable that is being changed by the experimenter is called the INDEPENDANT ______________________ variable. » The variable that changes as a result of the independent variable is called DEPENDENT the ________________ variable. » In physics there is no CONTROL variable, instead there are » CONTROLED variables is what is kept the same throughout the experiment because it also affects the dependent variables being tested, thus affecting the outcome of the experiment Your knowledge and grade in physics is related to the amount of quality time you spend studying.. independent variable: ? dependent variable: ? » If you are measuring the amount of bacteria over time, what is the independent and dependent variable? Dependent Variable on the y-axis! Independent Variable on the x-axis! The graph shown here is ok. Why only ok? The graph shows us the overall trend (average value). In science we care about BOTH the average value and some measure of the variation/uncertainty in the data. We use the synonymous terms uncertainty, error, or deviation to represent the variation in measured data. Data nearly always has uncertainty. There are many causes of uncertainty (different results when doing the same measurements), including: instrument uncertainty No measuring instrument is fully precise. Each instrument has an inherent amount of uncertainty in its measurement. Even the most precise measuring device cannot give the actual value because to do so would require an infinitely precise instrument. uncontrolled variables Its nearly impossible to fully control all variables. Un- or poorly- controlled variables will cause variation in your results. For example, changes in wind speed and direction would effect the results of a projectile launching experiment. Sampling from a population Sometimes you are measuring a small number of individuals from a larger population who differ in some trait. This is very common in biology but is NOT usually relevant to physics. Random and systematic errors » In ANY science, we care about the variation in the data nearly as much as the overall outcome / trend. Why? » Because the variation gives us some (NOT perfect) idea of how close our measurement is to the TRUE value. » Variation also limits our ability to make a conclusion about an overall trend or difference in our data. For example, our measurements of the size of 100 maple leaves might be summarized by reporting a typical value and a range of variation. This data can be reported in the form of a plot with "error bars". For example, if 100 maple leaves were collected from three different sites (parking lots, prairie, and the woods) we can display typical values and ranges of variation: As you can see there is a huge deviation from mean value for parking location, so it is not quite possible to distinguish between parking and prairie location. The uncertainty gives an answer to the question, "how well does the measurement represent the value of the quantity being measured?" In general: Measurement = (mean value ± uncertainty) unit of measurement uncertainty uncertainty mean value Measurement values Uncertainty (of measurement) defines a range of the values that represent the measurement result at a given confidence. Unfortunately, there is no general rule for determining the uncertainty. If STANDARD DEVIATION is taken as uncertainty then Measurement = (mean value ± standard deviation) indicates approximately a 68% confidence interval Usually done in fields where a sample represents the whole population (biology, sociology, psychology …) In PHYSICS, Chemistry there is no sample representing the whole population , so uncertainty is 𝑨𝒃𝒔𝒐𝒍𝒖𝒕𝒆 𝒖𝒏𝒄𝒆𝒓𝒕𝒂𝒊𝒏𝒕𝒚 ∆𝒙 = 𝑟𝑎𝑛𝑔𝑒 2 𝑹𝒂𝒏𝒈𝒆 𝒐𝒇 𝒕𝒉𝒆 𝒎𝒆𝒂𝒔𝒖𝒓𝒆𝒎𝒆𝒏𝒕𝒔 = 𝑥𝑚𝑎𝑥 − 𝑥𝑚𝑖𝑛 For example, result (20.1 cm ± 0.1) cm basically communicates that the person making the measurement believe the value to be closest to 20.1 cm but it could have been anywhere between 20.0 cm and 20.2 cm. A major part of any lab is evaluating the strength of your data, identifying the possible sources of error, and suggesting improvements or further lines of inquiry for the future. DO NOT ANSWER: “I think my data are good because I worked hard and didn’t mess up. I could do better if I had more time and better equipment.” Instead, » use your uncertainties to comment meaningfully on whether or not you can really make a conclusion or find a trend. » Think carefully about sources of error (often, these are poorly controlled variables that are affecting your results, but sometimes your whole approach may be flawed). » If your uncertainties are really large you should explain why and suggest how to reduce them in the future. » If you obtained the expected results then think of a follow up question you could study. If the results were surprising, think of a way to figure out why you got those results. Human error (messing up) is not a valid type of experimental error, thus the term “HUMAN ERROR“ should NEVER be used in a lab report!!! Instead the two acceptable types of experimental error explored in physics labs are systematic and random error. LABS ARE NOT MEANT TO BE FUN, OR EASY, OR TO BOOST YOUR GRADES. (though, I do hope you enjoy it, and I always always hope you earn a good grade!) We do labs so that you learn how to investigate a problem and how to analyze data. By 11th grade, you should know how to investigate a problem fairly well. One of our big focuses this year, then, is how to analyze data in a rigorous fashion. (so, use this ppt as a reference AND DO YOUR BEST WORK on the upcoming lab!!) 3 A REPORTING A SINGLE MEASUREMENT – Instrument uncertainty No measuring instrument is fully precise. Each instrument has an inherent amount of uncertainty in its measurement. Even the most precise measuring device cannot give the actual value because to do so would require an infinitely precise instrument. Generally we report the measured value of something with the decimal place or precision going not beyond the smallest graduation (called the ‘least count’) on the instrument. We report instrument uncertainty whenever we take a single measurement of something … If not stated differently by manufacturer, these uncertainties are: Analog instrument (the one with a scale): ½ of the smallest increment (precision) Digital instrument : the whole smallest increment Uncertainties are given to 1 significant figure. # of SF is 4: 3 known with certainty and one estimate L = (10.66 ± 0.05) cm The experimenter can determine the error to be different from instrument uncertainty provided some justification can be given. For example, mercury and alcohol thermometers are quite often not as accurate as the instrument uncertainty says. Instrument uncertainty when measuring time with stopwatch is certainly not the one stated by manufacturer – usually 0.01 s. It is ridiculous since you could never, ever move your thumb that fast! It doesn’t take into account human reaction time. More realistic would be ± 0.3 s , if you can justify it. Example: Distance of horizontal travel of a projectile launched at different angles. Angle of launch (+ 5o) Horizontal Distance travelled (m) (+ 1 cm) Trial 1 Trial 2 Trial 3 Trial 4 Trial 5 20 6.2 6.6 6.7 6.2 6.4 40 9.9 9.6 9.5 10.1 9.8 60 8.7 9.1 8.6 8.3 8.6 80 3.1 3.5 3.7 3.2 3.3 When someone else reads this report they would know instrument uncertainties. If you add the sentence like: Although measurement tape had instrument uncertainty of 0.5 mm, due to set up of experiment (….) estimated uncertainty is 1 cm (reason). 3 B REPORTING YOUR BEST ESTIMATE OF A MEASUREMENT – Absolute uncertainty The best way to come up with a good measurement of something is to take several measurements and average them all together. Each individual measurement has uncertainty, but the reported uncertainty in your average value is different than the uncertainty in your instrument. The precision of the instrument is not the same as the uncertainty in the measurement (unless you take only one measurement). Taking several measurements of something (x) , leads to distribution of values (x1, x2, x3,…) 𝑨𝒗𝒆𝒓𝒂𝒈𝒆 𝒗𝒂𝒍𝒖𝒆 = 𝑠𝑢𝑚 𝑜𝑓 𝑎𝑙𝑙 𝑚𝑒𝑎𝑠𝑢𝑟𝑒𝑚𝑒𝑛𝑡𝑠 : 𝑛𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑚𝑒𝑎𝑠𝑢𝑟𝑒𝑚𝑒𝑛𝑡𝑠 𝑥𝑎𝑣𝑔 = 𝑥𝑖 𝑛 𝑹𝒂𝒏𝒈𝒆 𝒐𝒇 𝒕𝒉𝒆 𝒎𝒆𝒂𝒔𝒖𝒓𝒆𝒎𝒆𝒏𝒕𝒔 = 𝑀𝑎𝑥𝑖𝑚𝑢𝑚 𝑣𝑎𝑙𝑢𝑒 − 𝑚𝑖𝑛𝑖𝑚𝑢𝑚 𝑣𝑎𝑙𝑢𝑒 = 𝑥𝑚𝑎𝑥 − 𝑥𝑚𝑖𝑛 𝑨𝒃𝒔𝒐𝒍𝒖𝒕𝒆 𝒖𝒏𝒄𝒆𝒓𝒕𝒂𝒊𝒏𝒕𝒚 ∆𝒙 = 𝑟𝑎𝑛𝑔𝑒 𝑥𝑚𝑎𝑥 − 𝑥𝑚𝑖𝑛 = 2 2 𝑭𝒓𝒂𝒄𝒕𝒊𝒐𝒏𝒂𝒍 𝒖𝒏𝒄𝒆𝒓𝒕𝒂𝒊𝒏𝒕𝒚 = ± ∆𝑥 𝑥 𝑷𝒆𝒓𝒄𝒆𝒏𝒕𝒂𝒈𝒆 𝒖𝒏𝒄𝒆𝒓𝒕𝒂𝒊𝒏𝒕𝒊𝒆𝒔 = ± (𝑥 𝑖𝑠 𝑎𝑐𝑐𝑡𝑢𝑎𝑙𝑙𝑦 𝑥𝑎𝑣𝑔 ) ∆𝑥 100% 𝑥 (𝑥 𝑖𝑠 𝑎𝑐𝑐𝑡𝑢𝑎𝑙𝑙𝑦 𝑥𝑎𝑣𝑔 ) Measurement = average value ± absolute uncertainty (unit of measurement) 𝑥 = (𝑥𝑎𝑣𝑔 ±∆𝑥) 𝑢𝑛𝑖𝑡𝑠 Always round your stated uncertainty up to match the number of decimal places of your measurement, if necessary. In the IB Physics laboratory, you should take 5 measurements if you can manage it. EX: The Six students measure the resistance of a lamp. Their answers in Ω are: 609; 666; 639; 661; 654; 628. What should the students reports as the resistance of the lamp? Average resistance = 643 Ω Range = Largest - smallest resistance: 666 - 609 = 57 Ω Absolute uncertainty: dividing the range by 2 = 29 Ω So, the resistance of the lamp is reported as: R = (643 ± 30) Ω 𝐹𝑟𝑎𝑐𝑡𝑖𝑜𝑛𝑎𝑙 𝑢𝑛𝑐𝑒𝑟𝑡𝑎𝑖𝑛𝑡𝑦 = ±0.05 𝑃𝑒𝑟𝑐𝑒𝑛𝑡𝑎𝑔𝑒 𝑢𝑛𝑐𝑒𝑟𝑡𝑎𝑖𝑛𝑡𝑖𝑒𝑠 = 5% This tells you immediately the maximum and minimum experimental values of a measurement When taking time measurements, the stated uncertainty cannot be unreasonably small – not smaller than 0.3 s, no matter what the range. For example if the range is 0.1s, then absolute uncertainty is 0.05 s which is highly unreasonable. You have to change it and explain the reasons. When taking several measurements, it should be clear if you have a value with a large error. Do not be afraid to throw out any measurement that is clearly a mistake. You will never be penalized for this if you explain your rationale for doing so. In fact, it is permissible, if you have many measurements, to throw out the maximum and minimum values. Remember that the manner in which you report measured and calculated values is entirely up to you, as the experimenter. However, be realistic in your precision and be able to fully justify reported measurements and calculated values, showing all of your work in doing so. 4. DRAWING A GRAPH In many cases, the best way to present and analyze data is to make a graph. A graph is a visual representation of 2 things and shows nicely how they are related. A graph is the visual display of quantitative information and allows us to recognize trends in data. Graphs also let you display uncertainties nicely. When making graphs: 1. The independent variable is on the x-axis and the dependent variable is on the y-axis. 2. Every graph should have a title that this concise but descriptive, in the form ‘Graph of (dependent variable) vs. (independent variable)’. 3. The scales of the axes should suit the data ranges. 4. The axes should be labeled with the variable, units, and instrument uncertainties. 5. The data points should be clear. 7. Error bars should be shown correctly (using a straight-edge). Error bars represent the uncertainty range 8. Data points (average value ONLY!) should not be connected dot-to-dot fashion. A line of best fit should be drawn instead. The best fit line is not necessarily the straight line and should pass through all of the crosses created by the error bars. Approximately the same number of data points should be above your line as below it. 9. Each point that does not fit with the best fit line should be identified. If you have an outlier, leave it as it is, discuss it, and the draw a second graph omitting the outlier and discuss it again. 10. Think about whether the origin should be included in your graph (what is the physical significance of that point?) Do not assume that the line should pass through the origin. A nice way to show uncertainty in data is with error bars. These are bars in the x and y directions around each data point that show immediately how big or small the uncertainty is for that value. Plotting a graph allows one to visualize all the readings at one time. Ideally all of the points should be plotted with their error bars. In principle, the size of the error bar could well be different for every single point and so they should be individually worked out. In practice, it would often take too much time to add much time to add all the correct error bars, so some (or all) of the following short cut could be considered. ▪ Rather than working out error bars for each point – use the worst value and assume that all of the other errors bars are the same. In the report include explanation, a sentence like: “Taking the highest uncertainty we are reasonably sure that result is ….” You should discuss the experiment results at the end. What type of dependence you discovered,…. For example when graphing experimental data, you can see immediately if you are dealing with random or systematic errors (if you can compare with theoretical or expected results).