Civil Group 3 Official White Paper As Prepared for the Probability and Statistics Engineering Course Prepared By: Robert Ruppel Cole Thorton Andre Cozier Yunior Gomez Nieto Daniel Avalos-Medina 1 Table of Contents Introduction …...................................................................................3 Case 1: Risky Business in the Construction World........................4-8 Case 2: Using Artificial Neural Networks to Calculate Significance of Parameters................................................................................9-13 Case 3: Review of a Performance Analysis of Predicting People`s Presence in a Building................................................................14-17 Case 4: Using Limit Equilibrium Methods to Recalculate the Factor of Safety for Slope Failures........................................................18-23 Case 5: Review of Correlative Analysis of Steel Characteristics.............................................................................24-28 Conclusion.......................................................................................28 References........................................................................................29 2 Introduction As a valued client, the full understanding, implementation and transparency of projects past, present, and future are key elements to which our team is committed to serving you. Among the several values such as honesty, reliability, and loyalty that our team upholds, we put safety above all. With safety as our corner stone, our team can then put full effort into completing our client’s demands with full confidence that hazard is mitigated. The development of infrastructure in dense and populated areas follow strict guidelines as to uphold the safety of citizens. Our team makes safety in all aspects of the construction and civil engineering discipline a top priority to ensure longevity in our partnerships. Motivated by our successes in past projects, we are excited to present the following cases which demonstrate our team’s emphasis on safety and reliability by implementing a variety of statistical methods. This safety starts at the foundations of these large projects to which we analyze the slope failure of soils using descriptive statistical methods to yield higher factors of safety. From then, the implementation of Gauss, Lognormal and Shift Lognorm probability density functions help our team choose materials that hold up against their appropriate loads while modeling data of yield strength and partial safety factor elements. Soon after, water utilities must be placed. By using and analyzing water quality data, we estimate potential hazards in the water resources that may be below quality standards. The Monte Carlo method as well as the Artificial Neural Network gives our team the ability to identify factors that influence chlorine decay in water, subsequently leading to health and hazard alleviation for occupants of the building. This safety protocol extends to our construction workers and other building trades as our team hires and contracts a variety of building trades to complete the project. The use of statistical methods such as risk matrices and risk analyses allow our team to prevent fatal and non-fatal work-related injuries among our hired building trades such as carpenters, electricians, and equipment operators on site. To maximize efficiency among these many moving trades, the use of inferential statistics and probability models are applied to infer occupancy of roads, equipment and rooms. By avoiding redundancy in tasks and maximizing the use of time to complete projects, costs of operation go down, saving both the company and the client money. 3 Risky Business in the Construction World By: Andre Cozier This case study reviews the article Comparative Injury and Fatality Risk Analysis of Building Trades by Selim Baradan1 and Mumtaz A. Usmen. Study Objective The construction industry has many assumptions about the life-jeopardizing risk of building trades. This study uses data from the Bureau of Labor Statistics on injury and fatalities among all industries under the construction of buildings to develop a coherent understanding of what specific trades are at most risk. There will be 16 trades scrutinized in this article relating to the construction of buildings. Study Outcome Out of the article’s focus on 16 trades, the workers with the highest fatality rates are roofers with a score of 55.30 with iron workers in second highest risk with a score of 52.70. The least risky building related trades that have a non-zero risk score are equipment operators, however, the trades with no risk score (no measured occupational related deaths in a given year) are carpet, floor, and tile installers and finishers, cement masons, concrete and terrazzo finishers, glaziers, plasterers and stucco masons, tile setters, and marble setters. These scores will be further understood in the following ‘Data Used for the Research’ section below. Data Used for the Research The article is largely based on data collected in a 2003 published article by the Bureau of Labor Statistics to provide information on risk in construction related jobs. Unfortunately, this data has had negligible changes in the decade preceding the study with an average of 1,126 fatalities in the construction industry per year. The data used by the BLS (Bureau of Labor Statistics) in this study includes annual incidence and fatality rates. The incidence rate is a standard measurement calculated by the number of injuries or illnesses from 100 workers in a specific field. These 100 workers are assumed to be full time workers based off a 40-hour work week for 50 weeks in a year. The data from these 100 workers per each field are then used as a microcosm for the entirety of their respective industry in the United States. The fatality rate is then defined as the number of occupational deaths per 100,000 workers in a year and is used to represent the deaths in the respective industry per year. Among data relating to the risk of fatalities by the 16 trades listed (iron workers, roofers, carpet, floor, tile installers, drywall installers, insulation workers, carpenters, brick masons, plumbers, electricians, plasterers and stucco masons, tile setters, and marble setters, glaziers, painters and paperhangers, sheet metal workers, cement masons, concrete finishers, and construction equipment operators), the study analyzes the cost of time lost after a non-fatal injury, median wages among each of the 16 trades, and the median amount of days away from work to develop a final risk assessment score. 4 Table 1: Risk of Construction Related Jobs The risk assessment score begins with a chart that ranks the index of risk from 0.01 to 1.5 or more in roughly 0.25 increments. The index is then correlated to a score from 1-7, 7 being the riskiest. It is noted that the index numbers that signify risks for a certain trade or job found in the North American Industry Classification System. The table below shows the Risk Score Criteria. Table 2: Fatalities vs Nonfatal Injuries The risk score is calculated by multiplying the risk score of the fatalities by its respective index number and adding it to the risk score of nonfatal injuries. 5 Statistical Analysis Review Inferential statistics is mostly used in the collection of data by the BLS, for the data of 100 workers is a representation of a larger population in the United States. This can also be seen in the 16 trades being represented. With respect to the methodology used, risk analysis is the overarching tool used for developing the figures. Risk analysis in an overarching technique that applies largely to fields where physical risk is inevitable. This method is typically used during the design and development phases of a variety of system safety applications, including chemical processing, manufacturing, automobile, cruise, and plane industries for assessing the risk associated with different scenarios. Furthermore, risk analysis methodologies implement risk control measures to mitigate the likelihood of damage if a certain risk value is unacceptable. A risk matrix is then a technique under risk analysis which uses the extent of damages and the likelihood of occurrence to describe a specific event. The occurrences are measured as fifths of 100 (0-20%, 21-40%, 41-60%, 61-80%, and 81-100%) and labelled in respective columns as “impossible”, “unlikely”, “possible”, “likely”, and “highly likely”. The article further extended the concepts of risk analysis and risk matrices to occupational safety in construction related fields. This was done by the implementation of inferential statistical data from 16 building trades. Evolution of the Article’s Text While the article is targeted towards occupants of building related trades, the study effectively delivers the information in colloquial English as to inform any person interested in such topic. Despite dealing with very specific statistical methodologies, the study synopsizes the data extruded from the BLS to produce a legible article which almost anyone can understand. Presenting the construction industry and its various subsequent trades in the introduction leads the reader to understand what the article will be focusing on. The article then flows toward a more analytical form of communication in the context of risk score criteria and risk ranking. Following the explanation of methodologies and graphs, the summary and conclusion accomplish its goal in highlighting the study’s takeaways. This section is the most comprehensible as it reviews the content chronologically and focuses on the importance of understanding risk analysis using quantitative techniques. Evaluation of the Articles Graphics The use of tables and graphs are prevalent in the study. The tables in the article include two charts which inform the reader on the data being analyzed. The graphs are then dependent on the two tables. These two graphs include one line graph that correlates nonfatal injuries with the cost of lost time. This line graph is shown below. 6 Table 3: Nonfatal Injury Rate vs Cost of Lost Time The next graph is a bar graph that compares nonfatal injuries, fatalities, and the combination of the fatal and nonfatal injures with building trades and their risk score. This bar graph is shown below. Table 4: Building Trades and Risk Scores The graphs are important in demonstrating the findings of the study without much prior 7 Contribution of the Research The study mentions their sources toward the end of the article. Out of the 16 sources mentioned, the Bureau of Labor Statistics and the Department of Health and Human Services seem to be the most commonly used within the in-text citations. The article is consistent in giving credit to the suppliers of the data, as well as formulating conclusions from said data on their own merit, 8 Using Artificial Neural Networks to Calculate Significance of Parameters By: Daniel Avalos-Medina This case study reviews the article Using artificial neural network models to assess water quality in water distribution networks by G.A Cuesta Cordoba, L. Tuhovacak, and M. Taus Study Objective The purpose of this research was to use a model to simulate chlorine decay under a specific zone of the Water Distribution System in the district of Brno-Kohoutovice, Czech Republic. The Artificial Neural Network along with the Monte Carlo method were both developed to simulate chlorine decay at selected nodes with good accuracy while determining its influencing factors. Study Outcome The key parameters to chlorine decay are initial chlorine, temperature and flow. These parameters have the largest influence in predicting the chlorine decay at pressure zone 1.3.1 Kohoutovice. Temperature at different location at the zone should be ideally calibrated to consider the reaction of chorine at different temperatures during different seasons of the year. The larger the database containing historic data and parameters, the more capable Artificial Neural Network is to be precise in conducting the results of the simulation. Data Used for the Research The data was collected by a water system that is operated by Brnenske vodarny a kanalizace, (BVK) which is located in the city district of Brno-Kohoutovice in the country of Czech Republic. The BVK provides water to a population consisting of 13,338 people. Data was obtained from two tanks called Bosonohly and Kohoutovice, which provide water to the area with each one consisting of two chambers. The data consists of the geometric dimensions of the tanks and the maximum water level for each. Historical data of flow, pH, turbidity, and initial chlorine was collected to create a data base for the Artificial Neural Network to make predictions of the parameters of chlorine decay. Data collected included 667 values for parameters and was obtained by BVK during September and October of 2011. Any data missing was fulfilled by the Monte Carlo method to manage the level of uncertainty and provide more accurate results. 9 Statistical Analysis Review The researchers used two different types of methods to examine the influential factors causing chlorine decay in the water. The Artificial Neural Network method simulates the human brains capability of analyzing and processing information. This method uses historical data to make predictions of the free chlorine at Kohooutovice. The input parameters given by the database will be represented by numerical values in the input layer. The input layer sends information to hidden layers which is where values are added to identify if the input layer`s data have any significant components while simulating different possible outcomes. The Monte Carlo method is a complimentary method that is used to make sure the Artificial Neural Network provides accurate output results when having missing information from a data base. This method provides estimates of the variance and provides a confidence region that is used to formulate the missing data. The Artificial Neural Network has two limitations to its capabilities of simulating the outcomes of the experiment. The Artificial Neural Network is only as accurate as the historical data given to it. Care must be taken when the data is collected. The information requires verification and must be current. The second limitation of this method involves the lack of data that may be provided by a database. Missing historical data will affect the accuracy of the output provided by the Artificial Neural Network. The objective of this method is to use as much information given to the input layer to evaluate the combinations and weigh the parameter`s capability to influence the chlorine decay. The Monte Carlo method is used to produce an analytical value for the given missing data. It generates these values by following either a normal or uniform distribution depending of the simulated parameters. These generated values are then compared to actual values and analyzed with descriptive statistics to check the significance of the simulated parameters. After all the data is collected, the Artificial Neural Network may be plotted to calculate the chlorine decay. Based on analyzing the results, the researchers were able to identify that the initial chlorine, flow and temperature have the largest influence in chlorine decay. This was identified by the output layer generating the highest values for these three factors. These values were generated after going through all the hidden layers, consisting of algorithms calculating the probabilities and weighing factors. Evaluation of the Article`s Text The author`s article was written for the water utility industry. This industry`s goal is to provide water to consumers with a certain standard that meets the required continuity, pressure flow and water quality. This goal is difficult to achieve due to water distribution having many factors that may alter these standards such as the components of the Water Distribution System, for example the materials utilized to construct the valves, tanks, and pipes. The target audience was identified 10 by the article’s focus on how these methods could be used to preserve the quality of water in this industry. The technical terminology used to describe the components of each method and software included in this article can make it difficult to comprehend for someone who is not familiar with the field of statistics or within the water industry. This research can be of great use to those industries and businesses that may have various number of outcomes depending on multiple variables when dealing with large data sets in their respected field. In order to make this appealing and accessible to these different industries and businesses, a more descriptive way of setting up the input parameters for an Artificial Neural Network model would have to be provided. This could make the audience more aware of how they could implement this methodology in their business scheme prior to taking action in becoming more energy and time efficient. Evaluation of the Article`s Graphics This article is composed of many graphics and tables that help the audience understand the procedures that are taken place when simulation is occurring. The tables are well structured and implement the role of descriptive statistics prior to the test being conducted, such as its mean, standard deviation, maximum, and minimum. Figure 1 demonstrates how the input parameters leads to an output by going across the hidden layer, while only showing one hidden layer for the sake of comprehension with proper labeling. Table 1 does a good job of expressing the descriptive variables taken into account when the Monte Carlo calculations are generated. Even though the information itself is very insightful, what they are trying to accomplish could be expressed clearer by breaking down the table into multiple tables. The current table holds too much information that may confuse the audience or be considered overwhelming. This could result in the audience not wanting to take the time to analyze the table. All graphics and tables are found directly below the paragraph from which it was referenced by. This is helpful by guiding the audience through a complicated experiment and determining what particular portions of the data is being evaluated. 11 Figure 1: Artificial Neural Network computation illustration Table 5: Statistics of the parameters calculated by Monte Carlo method Contribution of the Research This research provides a different way of collecting information and evaluating conditions without having to physically conduct the experiment again to retrieve the data. Certain data from water treatment plants require a large amount of data to be collected. With Artificial Neural Network the researchers can run a simulation that can predict outcomes based on historical data. This allows companies in the water quality industry a more time efficient way of collecting accurate data and weighing the influential factors that provide an output. As for the research itself, the research showed that key parameters to chlorine decay is the initial chlorine, flow, and temperature. These results are critical when attempting to distribute water at a large scale while 12 maintaining the quality standards and features of the water. New advancements in material of pipes may be developed to regulate the flow and the temperature of the water within the pipes. The Artificial Neural Network can play a great role in structural engineering. The Artificial Neural Network can be used when calculating certain maximum strength and hardness of a material. As an engineer, it is important to be knowledgeable about the material that is being use for a particular structure to avoid failure. For engineers to find certain characteristics about a material they typically take a sample of the material and test it by using a particular test, like the tension test and the Rockwell Hardness test. These tests usually require multiple trials and many disposable materials to run the experiment on. The Artificial Neural Network can use historic data from the material in question to generate an accurate output calculation. This method may also be used to evaluate what characteristics of a particular material are most influential when having a certain property. Statistics and probability provide an efficient, accurate, and less wasteful way of collecting information in the field of structural engineering. 13 Review of a Performance Analysis of Predicting People’s Presence in a Building By: Yunior Gomez Nieto This case study reviews the article Predicting people’s presence in buildings: An empirically based model performance analysis by Ardeshir Mahdavi and Farhang Tahmasebi. Figure 2: Outgoing refurbishment of a building In the University of Vienna Source: TU Wien Study Objective Knowing future presence is of great importance in implementing building systems control strategies that are dependent on the occupant’s interaction with systems such as windows, fans, lights, and radiators. That is why the study choose to evaluate several occupancy models and explore the potential of predicting the future presence of occupants. To quickly summarize the study; it evaluates two existing probabilistic occupancy models, the Reinhart and the Page, and runs them on the same occupancy data. Then to distinguish how well the models performed in predicting daily occupancy profiles, they were individually compared to a simple non-probabilistic model. Study Outcome After running the three models on the occupancy data to predict whether a room is occupied or vacant, the resulting predictions and the actually monitored occupancy profiles were compared. This comparison simply showed whether the models were correct or wrong, but out of it, five statistics were chosen to analyze the model’s performance towards each other. These statistics all showed errors in predicted arrival time, departure time, duration of occupancy, occupancy state on a daily basis, and a number of daily transitions of going from occupied to vacant. The statistics were then turned into cumulative distribution functions where they can be better visualized as graphs. 14 The study found out that the level of predictive accuracy was low for all the models and admits that further using them in more practical field applications would result in even poorer results. This can be attributed to having the predictive model trained for only one location which is an ideal situation that’s unlikely to be presented in practical use. Figure 3: Cumulative Distribution Functions for Statistic Arrival Time Source: Energy and Building 15 Data Used for the Research The occupancy data was collected from one of the office buildings at the Vienna University of Technology. Wireless ceiling-mounted motion sensors were placed in eight workplaces and gave a value of 0 if vacant and a value of 1 if occupied. The sensors also came included with the abilities to note the duration of either event and labeling the dominant occupancy state for an interval. Data such as the time before 8 AM and after 7:45 PM and nonworking days such as business trips, sickness, and holidays were not included as part of the study since they didn’t pertain to the study’s objective. The occupancy data totaled 90 working days for a single run, but to have a wider range of results, the study opted to use a Monte Carlos algorithm. This program created 100 simulations, increasing the amount of available data that can be run through the models. Figure 4: Sample Binary Occupancy Profile Generated by the MT Model Source: Energy and Building Statistical Analysis Review The data collected from the University is referred as bivariate since it comes with only two possibilities: either occupied or vacant. This type of data is qualitative in nature since it can’t be expressed in numbers, but once the study made the number one represent occupied and zero represent vacancy, it became quantitative. This binary form of data belongs to the descriptive branch of statistics where random variables can only occur in a finite interval and are used to describe features in a study such as mean, standard deviation, and variance. When the data was run through the models, it was turned into predictive data in the form of a best-fitting threshold graph. Aside from visually presenting how likely there is to be occupancy at a certain time, it also has a threshold value that suggests an acceptable chance in which the probability will occur. This is no longer descriptive statistics, but inferential since it’s used to deduct whether occupancy would happen from a sample population to a bigger one. The predictive data was then compared to the actual monitored data which lead to its final stage. The result was five descriptive statistics that “summarize important properties of basic data’’, and when converted into cumulative distribution functions showed how well was the performance for each model. The stages of the data from the beginning to the end were to get it from “a readable and worthwhile form’’ into a point where it would “allow us to obtain a graphical representation of the data”, therefore meeting the study goal of representing the performance of the models against each other. 16 Evaluation of the Article’s Text The article is written towards an audience that has a good understanding of statistical concepts. With technical language that vastly makes up the study, the little or no information that goes towards explaining the processes and statistical concepts, and especially a topic that isn’t geared for beginners, someone without any experience or even with a small background in it would find it difficult to grasp. With how it’s currently written, it can be argued that it’s unlikely that an inexperienced audience would make any use of this reading, but by going more in-depth into the process rather than merely stating what they did, using simpler vocabulary and making a section defining what can’t be replaced, it would let the reader make more sense of the work. Evaluation of the Article’s Graphics The article consists of three types of graphs that serve to illustrate the resulting progress of their work and aren’t necessary for the understanding of the study. The first type of graph, the cumulative distribution function, is used to show the difference in predictive performance between the three models based on five statistics. The bigger the function the more effective it is in its prediction. A negative aspect of this graph would be that it would be highly hard to comprehend without knowing what a CDF is and therefore inexperienced people wouldn’t understand its message. This goes hand in hand with my previous suggestion that there should be a more in-depth discussion on what is being said and, in this case, shown. The other type of graph is a histogram that shows the probability of occupation in two-hour intervals for 24 hours. As simple as it is, it is made more difficult with the addition of the best-fitting threshold, which easy to understand if you know what it is, but hard to find out what it even means with the little clarity provided and almost no trace definition on the internet. Without knowing what this line means, you miss out on essential information that needs to be grasped before continuing to the next graph. I again suggest adding a clarifying description to make it more understandable. The next and final graphic, the binary linear graph, shows the previous graph’s hours that meet the threshold and is based on occupancy state rather than probability presence. All of the graphs are positioned and sized perfectly and can be accessed smoothly after reading the information that pertains to it. Contribution of the Research This research presents towards the statistical modeling field that it’s not necessary to base your model on stochastic elements but, in this case, it more efficient to use non-probabilistic models rather than those that take into account the random diversity in accounting patterns. This contribution is useful in my field of civil engineering and any that revolve around the same area because it presents us with the most efficient model at this time. A model that can be used not for only this studies’ scenario of an office building but wider amount locations such as roads, parks, forests, and much more. This study demonstrates that probability/statistics can be of great use when designing many locations around the concept of occupancy. 17 Using Limit Equilibrium Methods to Recalculate the Factor of Safety for Slope Failures By: Cole Thornton This case study reviews the article Statistics for the Calculated Safety Factors of Undrained Failure Slopes by Erly Bahsan, Hung-Jiun Liao, Jianye Ching, and Szu-Wei Lee from Engineering Geology. Study Objective: The purpose of this article is to use the uncertainty of calculated factors of safety (FS) of soil slopes to create a relationship between the probability of failure of a man-made undrained clay slope and its calculated factor of safety to make the reliability design and reliability analysis easier. Study Outcome: The outcome of this report is that the two-dimensional limit equilibrium methods used in this article give unbiased estimates for calculating the factor of safety, that man-made soil slopes have a lower variability than natural soil slopes, and that the conversion of the tested shear strength value to the field shear strength value is an important part in the safety factor calculations. Data used for the Research: The data used in the article is from 43 previous cases found in literature, listed in the article, dating from 1956 to 2002 where the slope of clays have failed. All of these cases vary in the types of slopes used whether its natural, fill, or cut slopes. The purpose of reopening these cases was to recalculate the factor of safety for each case using both two-dimensional limit equilibrium methods known as Bishop’s Method and Spencer’s Method. Statistical Analysis Review: One of the main statistical methods used in the article is descriptive statistics, specifically variability. An example of variability within the article is the confirmation that natural failure slopes are much greater than man-made failure slopes when using the natural log of the factor of safety. Simplistic terminology like standard deviation and mean also come into play when dealing with human error in the analysis process for example. 18 A majority of the article uses descriptive statistics to analyze the differences between natural slopes and man-made slopes when calculating the natural log of the factor of safety for both. The article utilizes the term variability when using descriptive statistics to find correlations between natural slopes in three dimensions and man-made slopes in two dimensions since both act different under various circumstances. Variability is also used for finding more fundamental, statistical calculations like standard deviation and the mean of the magnitude of human error or the natural log of the factor of safety for both types of slopes. A likely reason why variability was conducted throughout the article is because of the possibilities descriptive statistics has when finding a number of possible outcomes as to why a slope fails or what the factor of safety is for either slope given certain parameters. The creators of the article most likely learned that utilizing descriptive statistics to its fullest potential can benefit in their answers and calculations for data that failed and never succeeded. Using standard deviation and the mean in various ways yields conclusive data that can be better narrowed down to how man-made slopes can be made more efficiently and how natural slopes can be taken with a more cautious approach. Evaluation of the Article’s Text: The presentation of the article is through the means of professionalism as well as through the perceptive of an engineer. The comprehensive and fairly complex text throughout the article conveys a higher understanding of the topics at hand while, at the same time, breaking down key phrases into digestible and understandable terminology for both engineers and non-engineers. Key phrases like “factors of safety” are phrases that engineers are familiar with which is why the article contains a quick definition to clarify the meaning for non-engineering audience members. For a majority of the article, however, the key points the article touches upon can be simplistic in nature when talking about the various slopes of soil and the mathematical methods used to achieve some of these values. The occupations the text in the article is directed towards are mostly geotechnical engineers, geologists, and professionals who specialize in various jobs surrounding soil and/or the environment. A niche topic like the various slopes of soil affecting the factor of safety are important to know for these professions but the topic may be trivial to some in another profession because of how minute the difference in slopes does to soils. A way this can be changed into a more understandable topic would be by tweaking the article to be more directed towards the dangers of the possibility of erosion and how the slopes of soil can be detrimental to lives of others. Tweaking the article to fit this agenda would give a whole new perspective to a different audience while keeping the same meaning behind the importance of various soil slopes, whether the slopes are man-made or natural. 19 Evaluation of the Article’s Graphics: Between the visualizations and the numerical graphs, there are multiple types of graphics being used for the information in the article. The first visualization is a two-dimensional representation of the profile of the slope. This contains a graph showing the specific water values for each layer of soil, the “x” and “y” axis values foe elevation and length, and the two methods that are used to calculate the slopes (i.e. Bishop’s Method and Spencer’s Method) that are found throughout the article. The second visualization is the profile of the borehole which contains the different layers of soil at the different elevations below the surface and what the shear strengths were at each elevation. The first set of graphs in the article are histograms indicating the frequency versus the factor of safety for both Bishop’s Method and Spencer’s Method. Both histograms also have a corresponding Q-Q plots with quantiles of input sample versus standard normal quantiles for both Bishop’s Method and Spencer’s Method. The other graph found in the article is a scatterplot with tread lines where the axes consist of the factor of safety by Spencer’s Method versus the factor of safety by Bishop’s Method. All of the visualization and numerical graphics are as follows: Table 6: Slope Profile of coordinate vs elevation 20 Table 7: Borehole Profile Table 8: Frequency Analysis 21 Table 9: Bishop`s method vs Spencer`s method The graphs and visualizations from the article are prepared in a way that they convey the message completely while still being readable and simplistic to understand. The information surrounding the graphics are very detailed in a mathematical and terminology sense with the equations neatly prepared and explained. The choice in which graphic to use is also appropriate and would not need changing. A reason for this is that some visualizations cannot be portrayed in any other manner like the boring log or the two-dimensional sideview of the slope profile so any other choice of visualization would not work as well. 22 Contribution to Research: This article proves that something as specific and simple as the slope of soil can have a major impact on the environment around us. Soil can be elusive when it comes to the long-term and short-term effects of building foundation above ground, whether the slope of the soil or the type of the soil is the factor. Depending on the conditions, the factor of safety comes into play with how secure the soil is and if the soil will be a problem or not. This is why research involving the information found in this article is so important to how future construction will take place. By having a better grasp on the understanding of soil slopes, one can create more secure foundations with higher factors of safety which allows for safer building construction when building on slanted surfaces or even flat surfaces. 23 Review of Correlative Analysis of Steel Characteristics By: Robert Ruppel This case study reviews the article Material and Geometrical Characteristics of Structural Steels Based on Statistical Analysis of Metallurgical Products. Study Objective The objective of this report was to evaluate the results of different strength measurements and characteristics into correlative relationships through the calculation and modeling of elements like yield strength and partial safety factors. Study Outcome The outcome of the report was the observation of the correlative behavior between the increase of steel plate thickness for any of the test steel types and the decrease in mean value yield strength of said steel plate. Data used for the Research This study was done through the analysis of already known material characteristics. That means that the data analyzed in the piece did not come from a specific experiment but rather was broken down through the variety of probability models explained in the Statistical Analysis Review portion. These inserted characteristics serve as the reports data and are cited as follows: • • • • • • • • • • • • • • • Mean Value of Yield Strength of steel s355 plates thickness ranges Standard Deviation of Yield Strength of steel s355 plates thickness ranges Skewness of Yield Strength of steel s355 plates thickness ranges Kurtosis of Yield Strength of steel s355 plates thickness ranges Gauss Distribution Value of steel s355 plates thickness ranges Logonormal Distribution Value of steel s355 plates thickness ranges Shift Logonorm Distribution Value of steel s355 plates thickness ranges Nominal Value of steel s355 plates thickness ranges Gauss Distribution Value for design value Rd Logonormal Distribution Value for design value Rd Shift Logonorm Distribution for design value Rd Design Value Gauss Distribution Value for partial safety factor Logonormal Distribution Value for partial safety factor Shift Logonorm Distribution for partial safety factor 24 • • • • Mean Value for section depths, web thicknesses, and flange thicknesses Standard Deviation for section depths, web thicknesses, and flange thicknesses Skewness for section depths, web thicknesses, and flange thicknesses Kurtosis for section depths, web thicknesses, and flange thicknesses Statistical Analysis Review This report was anchored in the fundamental elements of probability and statistics. This is made clear in the first section post the introduction where the first form of analysis gathered the mean, standard deviation, skewness and kurtosis of the yield strength per certain thickness ranges which are then utilized with the Monte Carlo method to figure number of measurements present and if they are enough to allow for accurate distribution functions being produced. This determination plays into the partial safety factor calculation as well. These elements play into the Hermite probability distribution function which is utilized over the Gauss density probability function due to its accounting for skewness and kurtosis. A real probability model is also created for the yield strength which is used to help define the nominal value for yield strength of certain thicknesses. Characteristic values for certain probability models are also present, being that of Gauss, Lognormal and Shift Lognorm probability density functions. All these elements played into creating the greater foundation of the Correlative Analysis Method utilized in this report. The Correlative Analysis Method was an ideal choice for the overall direction for this report. Correlative Analysis is a statistical method by which two quantitative variables are evaluated in terms of their connective relationship. This heavily present in the conclusion portion of the report that makes correlative statements upon the relationships of the quantities of mean yield strength and thickness. This correlation sets the grounds for the inferential statements later on with more correlations of the dimension tolerances and the tendency to go negative. Overall, the Correlative Analysis Method allowed for the greater correlation of mean yield strength lowering as thickness of the plate increases through the consolidation of all previous statistical elements into the statements of the correlative relationship stated in the conclusion. Evaluation of the Article’s Text This report is one of heavy detail and is delivered from a place of already built understanding. There are no portions of the piece that define any of the terms or methods utilized but rather takes any portion of explanation to tie in the terms/methods into the greater purpose of the piece. That tone and structure throughout the report makes it clear that this is prepared for an audience of professionals or academics already highly familiar with the know statistical and civil engineering elements of things like safety factors, standard deviations, probability models and yield strengths. The report does not seek to make the information hard to decipher and anyone with the proper background of understanding and knowledge would find the report to be more than appropriate. If the desire of the report’s creators was to reach a broader audience, then all that would be needed would be an additional portion of definitions and the addition of step by 25 step processes for some of the modelling. Outside of that the report succeeds in it purposes and is fit for the audience it seeks to reach. Evaluation of the Article’s Graphics The graphics utilized within the report fall into four categories. The first is the scatter plot graphs used to model correlating thickness values and thickness vs observations. Histograms were also used to model the reality frequency of certain dimension sections. Distribution plots were also utilized to mode the real probability of yield strength as well as the normal distribution showing the difference in Gauss and Hermite models. Finally, example figures of beam cross sections were utilized to showcase certain geometric characteristics being analyzed. Example images of the figures used are as follows: Figure 5: I Geometric Analysis Table 10: Cross-section flange Analysis 26 Table 11: Relative frequency vs section width Table 12: Probability vs Yield strength These graphics are expertly prepared, most likely formed from high level mathematic simulators. They are havinly detailed in their ploting and numbering. They are located in solid relation to where they are referenced in the report and are clearly labeled as to decrease any confusion. The only thing that could service a lesser informed individual would be some text boxes explaining some of the figures more directly instead of just referencing them. 27 Contribution of Research This study makes it clear that there is continuous need for monitoring of the observed characteristics of the steel samples as new designs come forward into the world. Though this report is not immediately intuitive to a layman, the methods present can be continuously conducted as time moves forward as to keep a continuous account of the state of the relationship between mean yield strength and thickness of steel. This is important for the idea of safety and repeated return to this report when designing a new structure would ensure that the safety of the structure and the reliability of the materials is properly managed. Conclusion The multiple cases studies presented in this paper highlights the variety of uses and impact probability and statistics has in the safety of our society. By allowing our team to do research in your USF’s college of civil engineering, we will be able to further explore and test the topics mentioned above as well as to the multitude of other fields that pertain to safety. Aside from the significant importance this research could have on the safety of our businesses, industry, and construction it would also be beneficial to the University as it would bring in grants and national recognition due to our work. Would you let our team carry out research on probability and statistics in the college of civil engineering? 28 References Zdeněk Kala, J. M. (2010, October 14). Material and geometrical characteristics of structural steels based on statistical analysis of metallurgical products. Journal of Civil Engineering and Management, pp. 299307. Mahdavi, A., & Tahmasebi, F. (2015). Predicting people's presence in buildings: An empirically based model performance analysis. Energy and Buildings, 86, 349-355. doi:10.1016/j.enbuild.2014.10.027 (2008) Descriptive Statistics. In: The Concise Encyclopedia of Statistics. Springer, New York, NY. https://doi-org.ezproxy.lib.usf.edu/10.1007/978-0-387-32833-1_108 Cordoba, G. C., Tuhovčák, L., & Tauš, M. (2014). Using Artificial Neural Network Models to Assess Water Quality in Water Distribution Networks. Procedia Engineering, 70, 399-408. doi:10.1016/j.proeng.2014.02.045 Baradan, S., & Usmen, M. A. (2006). Comparative Injury and Fatality Risk Analysis of Building Trades. Journal of Construction Engineering and Management, 132(5), 533–539. https://doi.org/10.1061/(asce)0733-9364(2006)132:5(533) Bureau of Labor Statistics. 2001. “BLS statistics on worker safety and health.” Bureau of Labor Statistics BLS http://bls.gov/bls/safety.htm May 25, 2005 Risk Analysis Methodologies | SSPA. (n.d.). Sspa.Se. Retrieved October 25, 2020, from https://www.sspa.se/tools-and-methods/risk-analysismethodologies#:%7E:text=Risk%20Analysis%20Methodologies%20are%20a,assessments%20of%20risk %20control%20measures. Risk Matrix. Visualize risks in a diagram. (n.d.). MicroTOOL. Retrieved October 24, 2020, from https://www.microtool.de/en/knowledge-base/what-is-a-risk-matrix/ Bahsan, Erly, et al. “Statistics for the Calculated Safety Factors of Undrained Failure Slopes.” Engineering Geology, vol. 172, 8 Apr. 2014, pp. 85–94., doi:10.1016/j.enggeo.2014.01.005. 29