{"id":226,"date":"2021-08-02T19:02:34","date_gmt":"2021-08-02T23:02:34","guid":{"rendered":"https:\/\/pressbooks.bccampus.ca\/statspsych\/?post_type=chapter&#038;p=226"},"modified":"2023-12-30T16:07:14","modified_gmt":"2023-12-30T21:07:14","slug":"chapter-4","status":"publish","type":"chapter","link":"https:\/\/pressbooks.bccampus.ca\/statspsych\/chapter\/chapter-4\/","title":{"raw":"4. Probability, Inferential Statistics, and Hypothesis Testing","rendered":"4. Probability, Inferential Statistics, and Hypothesis Testing"},"content":{"raw":"<h1><a id=\"4a\"><\/a>4a. Probability and Inferential Statistics<\/h1>\r\n<h6 style=\"text-align: right\"><a href=\"#video4a\">video lesson<\/a><\/h6>\r\nIn this chapter, we will focus on connecting concepts of [pb_glossary id=\"233\"]<strong>probability<\/strong> [\/pb_glossary] with the logic of inferential statistics.\r\n<span class=\"pullquote-right\">\"The whole problem with the world is that fools and fanatics are always so certain of themselves, and wiser people so full of doubts.\"\r\n\u2014 Bertrand Russel (1872-1970)<\/span>\r\n\r\nThese notable quotes represent why <strong>probability<\/strong> is critical for a basic understanding of scientific reasoning.\r\n\r\n<span class=\"pullquote-left\">\"Medicine is a science of uncertainty and an art of probability.\"\r\n\u2014 William Osler (1849\u20131919)<\/span>In many ways, the process of postsecondary education is all about instilling a sense of doubt and wonder, and the ability to estimate <strong>probabilities<\/strong>. As a matter of fact, that essentially sums up the entire reason why you are in this course. So let us tackle <strong>probability<\/strong>.\r\n\r\nWe will be keeping our coverage of <strong>probability<\/strong> to a very simple level, because the introductory statistics we will cover rely on only simple <strong>probability<\/strong>. That said, I encourage you to read further on compound and conditional <strong>probabilities<\/strong>, because they will certainly make you smarter at real-life decision making. We will briefly touch on examples of how bad people can be at using <strong>probability<\/strong> in real life, and we will then address what probability has to do with inferential statistics. Finally, I will introduce you to the <strong>[pb_glossary id=\"235\"]central limit theorem[\/pb_glossary]<\/strong>. This is probably one of the heftiest math concepts in the course, but worry not. Its implications are easy to learn, and the concepts behind it can be demonstrated empirically in the interactive exercises.\r\n\r\nFirst, we need to define <strong>probability<\/strong>. In a situation where several different outcomes are possible, the probability of any specific outcome is a fraction or proportion of all possible outcomes. Another way of saying that is this. If you wish to answer the question, \u201cWhat are the chances that outcome would have happened?\u201d, you can calculate the <strong>probability<\/strong> as the ratio of possible successful outcomes to all possible outcomes.\r\n<h4><a id=\"defineProb_return\"><\/a><a href=\"#defineProb\">Concept Practice: define probability<\/a><\/h4>\r\nPeople often use the rolling of dice as examples of simple <strong>probability<\/strong> problems.\r\n\r\n[caption id=\"\" align=\"aligncenter\" width=\"371\"]<img class=\"\" src=\"https:\/\/live.staticflickr.com\/62\/201651074_30058a36ac_b.jpg\" alt=\"Dice\" width=\"371\" height=\"278\" \/> \"Dice\" by matsuyuki is licensed under CC BY-SA 2.0[\/caption]\r\n\r\nIf you were to roll one typical die, which has a number on each side from 1 to 6, then the simple probability of rolling a 1 would be 1\/6. There are six possible outcomes, but only 1 of them is the successful outcome, that of rolling a 1.\r\n<h4><a id=\"calcProb1_return\"><\/a><a href=\"#calcProb1\">Concept Practice: calculate probability<\/a><\/h4>\r\nAnother common example used to introduce simple <strong>probability<\/strong> is cards. In a standard deck of casino cards, there are 52 cards. There are 4 aces in such a deck of cards (Aces are the \u201c1\u201d card, and there is 1 in each suit \u2013 hearts, spades, diamonds and clubs.)\r\n\r\n<img class=\"aligncenter wp-image-238 \" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.2-1024x576.png\" alt=\"\" width=\"539\" height=\"303\" \/>\r\n\r\nIf you were to ask the question \u201cwhat is the <strong>probability<\/strong> that a card drawn at random from a deck of cards will be an ace?\u201d, and you know all outcomes are equally likely, the <strong>probability<\/strong> would be the ratio of the number of times one could draw and ace divided by the number of all possible outcomes. In this example, then, the <strong>probability<\/strong> would be 4\/52. This ratio can be converted into a decimal: 4 divided by 52 is 0.077, or 7.7%. (Remember, to turn a decimal to a percent, you need to move the decimal place twice to the right.)\r\n<h4><a id=\"calcProb2_return\"><\/a><a href=\"#calcProb2\">Concept Practice: calculate probability<\/a><\/h4>\r\n<strong>Probability<\/strong> seems pretty straightforward, right? But people often misunderstand <strong>probability<\/strong> in real life. Take the idea of the lucky streak, for example. Let\u2019s say someone is rolling dice and they get 4 6\u2019s in a row. Lots of people might say that\u2019s a lucky streak and they might go as far as to say they should continue, because their luck is so good at the moment! According to the rules of <strong>probability<\/strong>, though, the next die roll has a 1\/6 chance of being a 6, just like all the others. True, the <strong>probability<\/strong> of a 4-in-a-row streak occurring is fairly slim: 1\/6 x 1\/6 x 1\/6 x 1\/6. But the fact is that this rare event does not predict future events (unless it is an unfair die!). Each time you roll a die, the <strong>probability<\/strong> of that event remains the same. That is what the human brain seems to have a really hard time accepting.\r\n<h4><a id=\"lucky_return\"><\/a><a href=\"#luckystreak\">Concept Practice: lucky streak<\/a><\/h4>\r\nWhen someone makes a prediction attached to a certain probability (e.g. there is only a 1% chance of an earthquake in the next week), and then that event occurs in spite of that low <strong>probability<\/strong> estimate (e.g. there is actually an earthquake the day after the prediction was made)\u2026 was that person wrong? No, not really, because they allowed for the possibility. Had they said there was a 0% chance, they would have been wrong.\r\n\r\n<strong>Probabilities<\/strong> are often used to express likelihood of outcomes under conditions of uncertainty. Like Bertrand Russell said, wise people rarely speak in terms of certainties. Because people so often misunderstand <strong>probability<\/strong>, or find irrational actions so hard to resist despite some understanding of <strong>probability<\/strong>, decision making in the realm of sciences needs to be designed to combat our natural human tendencies. What we are discussing now in terms of how to think about and calculate <strong>probabilities<\/strong> will form a core component of our decision-making framework as we move forward in the course.\r\n\r\nNow, let\u2019s take a look at how <strong>probability<\/strong> is used in statistics.\r\n<h4><a id=\"areaProb_return\"><\/a><a href=\"#areaProb\">Concept Practice: area under normal curve as probability<\/a><\/h4>\r\nWe saw that percentiles are expressions of area under a normal curve. Areas under the curve can be expressed as <strong>probability<\/strong>, too. For example, if we say the 50th percentile for IQ is 100, that can be expressed as: \"If I chose a person at random, there is a 50% chance that they will have an IQ score below 100.\u201d\r\n\r\n&nbsp;\r\n\r\n<img class=\"aligncenter wp-image-246 \" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.3-1024x553.jpg\" alt=\"\" width=\"534\" height=\"289\" \/>\r\n\r\nIf we find the 84th percentile for IQ is 115 there is another way to say that \u201cIf I chose a person at random, there is an 84% chance that they will have an IQ score below 115.\u201d\r\n\r\n<img class=\"aligncenter wp-image-245 \" style=\"font-size: 1em\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.4-1024x594.jpg\" alt=\"\" width=\"521\" height=\"302\" \/>\r\n<h4><a id=\"findPercentile_return\"><\/a><a href=\"#findPercentile\">Concept Practice: find percentiles<\/a><\/h4>\r\nAny time you are dealing with area under the normal curve, I encourage you to express that percentage in terms of <strong>probabilities<\/strong>. That will help you think clearly about what that area under the curve means once we get into the exercise of making decisions based on that information.\r\n<h4><a id=\"interpretPercentile_return\"><\/a><a href=\"#interpretPercentile\">Concept Practice: interpreting percentile as probability<\/a><\/h4>\r\n<strong>Probabilities<\/strong>, of course, range from 0 to 1 as proportions or fractions, and from 0% to 100% when expressed in percentage terms. In inferential statistics, we often express in terms of <strong>probability<\/strong> the likelihood that we would observe a particular score under a given normal curve model.\r\n<h4><a id=\"applyProb_return\"><\/a><a href=\"#applyProb\">Concept Practice: applying probability<\/a><\/h4>\r\nAlthough I encourage you to think of <strong>probabilities<\/strong> as percentages, the convention in statistics is to report to the probability of a score as a proportion, or decimal. The symbol used for \u201cprobability of score\u201d is <em>p<\/em>. In statistics, the interpretation of \u201c<em>p<\/em>\u201d is a delicate subject. Generations of researchers have been lazy in our understanding of what \u201c<em>p<\/em>\u201d: tells us, and we have tended to over-interpret this statistic. As we begin to work with \u201c<em>p<\/em>\u201d, I will ask you to memorize a mantra that will help you report its meaning accurately. For now, just keep in mind that most psychologists and psychology students <em>still<\/em> make mistakes in how they express and understand the meaning of \u201c<em>p<\/em>\u201d values. This will take time and effort to fix, but I am confident that your generation will learn to do better at a precise and careful understanding of what statistics like \u201c<em>p<\/em>\u201d tell us\u2026 and what they do not.\r\n\r\nTo give you a sense of what a statement of <em>p<\/em> &lt; .05 might mean, let us think back to our rat weights example.\r\n\r\n<img class=\"wp-image-201 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.10-1024x624.jpg\" alt=\"\" width=\"596\" height=\"363\" \/>\r\n\r\nIf I were to take a rat from our high-grain food group and place it on the distribution of untreated rat weights, and if it placed at <em>Z<\/em> = .9, we could look at the area under the curve from that point and above. That would tell us how likely it would be to observe such a heavy rat in the general population of nontreated rats -- those that eat a normal diet.\r\n\r\nThink of it this way. When we select a rat from our treatment group (those that ate the grain-heavy diet), and it is heavier than the average for a nontreated rat, there are two possible explanations for that observation. One is that the diet made him that way. As a scientist whose hypothesis is that a grain-heavy diet will make the rats weigh more, I\u2019m actually motivated to interpret the observation that way. I want to believe this event is meaningful, because it is consistent with my hypothesis! But the other possibility is that, by random chance, we picked a rat that was heavy to begin with. There are plenty of rats in the distribution of nontreated rats that were at least that heavy. So there is always some <strong>probability<\/strong> that we just randomly selected a heavier rat. In this case, if my treated rat\u2019s weight was less than one standard deviation above the mean, we saw in the chapter on normal curves that the <strong>probability<\/strong> of observing a rat weight that high or higher in the nontreated population was about 18%. That is not so unusual. It would not be terribly surprising if that outcome were simply the result of random chance rather than a result of the diet the rat had been eating.\r\n\r\nIf, on the other hand, the rat we measured was 2.5 standard deviations above the mean, the tail <strong>probability<\/strong> beyond that <strong>Z-score<\/strong> would be vanishingly small.\r\n\r\n<img class=\"aligncenter wp-image-202 \" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.11-1024x645.jpg\" alt=\"\" width=\"605\" height=\"381\" \/>\r\n\r\nThe <strong>probability<\/strong> of observing such a rat weight in the nontreated population is very low, so it is far less likely that observation can be accounted for just by random chance alone. As we accumulate more evidence, the <strong>probability<\/strong> they could have come at random from the nontreated <strong>[pb_glossary id=\"248\"]population[\/pb_glossary]<\/strong> will weigh into our decision making about whether the grain-heavy diet indeed causes rats to become heavier. This is the way <strong>probabilities<\/strong> are used in the process of <strong>[pb_glossary id=\"259\"]hypothesis testing[\/pb_glossary]<\/strong>, the logic of inferential statistics that we will look at soon.\r\n<h4><a id=\"statsProb_return\"><\/a><a href=\"#statsProb\">Concept Practice: statistics as probability<\/a><\/h4>\r\nNow that you have seen the relevance of <strong>probability<\/strong> to the decision making process that comprises inferential statistics, we have one more major learning objective: to become familiar with the <strong>central limit theorem<\/strong>.\r\n\r\nHowever, before we get to the <strong>central limit theorem<\/strong>, we need to be clear on the distinction between two concepts:\u00a0 <strong>[pb_glossary id=\"249\"]sample[\/pb_glossary]<\/strong> and <strong>[pb_glossary id=\"248\"]population[\/pb_glossary]<\/strong>. In the world of statistics, the <strong>population<\/strong> is defined as all possible individuals or scores about which we would ideally draw conclusions. When we refer to the characteristics, or parameters, that describe a <strong>population<\/strong>, we will use Greek letters.\u00a0A <strong>sample<\/strong> is defined as the individuals or scores about which we are actually drawing conclusions. When we refer to the characteristics, or statistics, that describe a <strong>sample<\/strong>, we will use English letters.\r\n\r\nIt is important to understand the difference between a <strong>population<\/strong> and a <strong>sample<\/strong>, and how they relate to one another, in order to comprehend the <strong>central limit theorem<\/strong> and its usefulness for statistics. From a <strong>population<\/strong> we can draw multiple <strong>samples<\/strong>. The larger <strong>sample<\/strong>, the more closely our <strong>sample<\/strong> will represent the <strong>population<\/strong>.\r\n\r\nThink of a Venn diagram.\u00a0<span style=\"text-align: initial;font-size: 1em\">There is a circle that is a <\/span><strong style=\"text-align: initial;font-size: 1em\">population<\/strong><span style=\"text-align: initial;font-size: 1em\">. Inside that large circle, you can draw an infinite number of smaller circles, each of which represents a <\/span><strong style=\"text-align: initial;font-size: 1em\">sample<\/strong><span style=\"text-align: initial;font-size: 1em\">.<\/span>\r\n\r\n<img class=\"size-medium wp-image-254 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.6-300x285.jpg\" alt=\"\" width=\"300\" height=\"285\" \/>\r\n\r\nThe larger that inner circle, the more of the <strong>population<\/strong> it contains, and thus the more representative it is.\r\n\r\n<img class=\"size-medium wp-image-253 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.5-300x283.jpg\" alt=\"\" width=\"300\" height=\"283\" \/>\r\n\r\nLet us take a concrete example. A <strong>population<\/strong> might be the depression screening scores for all current postsecondary students in Canada. A sample from that <strong>population<\/strong> might be depression screening scores for 500 randomly selected postsecondary students from several institutions across Canada. That seems a more reasonable proportion of the two million students in the <strong>population<\/strong> than a <strong>sample<\/strong> that contains only 5 students. The 500 student <strong>sample<\/strong> has a better shot at adequately representing the entire <strong>population<\/strong> than does the 5 student <strong>sample<\/strong>, right? You can see that intuitively\u2026 and once you learn the <strong>central limit theorem<\/strong>, you will see the mathematical demonstration of the importance of <strong>sample<\/strong> size for representing the <strong>population<\/strong>.\r\n\r\nTo conduct the inferential statistics we are using in this course, we will be using the normal curve model to estimate <strong>probabilities<\/strong> associated with particular scores. To do that, we need to assume that data are normally distributed. However, in real life, our data are almost never actually a perfect match for the normal curve.\r\n\r\nSo how can we reasonably make the normality assumption? Here\u2019s the thing. The <strong>central limit theorem<\/strong> is a mathematical principle that assures us that the normality assumption is a reasonable one as long as we have a decent <strong>sample<\/strong> size.\r\n\r\n<img class=\"alignright wp-image-243 \" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.1-353x1024.jpg\" alt=\"\" width=\"298\" height=\"864\" \/>\r\n\r\nAccording to the theorem, as long as we take a decent-sized <strong>sample<\/strong>, if we took many <strong>samples<\/strong> (10,000) of large enough size (30+) and took the mean each time, the distribution of those means w<span style=\"text-align: initial;font-size: 1em\">ill approach a normal distribution, even if the scores from each <\/span><strong style=\"text-align: initial;font-size: 1em\">sample<\/strong><span style=\"text-align: initial;font-size: 1em\"> are not normally distributed. To see this for yourself, take a look at the histograms shown on the right. The top histogram came from taking from a <\/span><strong style=\"text-align: initial;font-size: 1em\">population<\/strong><span style=\"text-align: initial;font-size: 1em\"> 10,000 <\/span><strong style=\"text-align: initial;font-size: 1em\">samples <\/strong><span style=\"text-align: initial;font-size: 1em\">of just one score each, and plotting them on a histogram. See how it has a flat, or rectangular shape? No way we could call that a shape approximating a normal<\/span><span style=\"text-align: initial;font-size: 1em\">\u00a0curve. Next is a histogram that came from taking the means of 10,000 <strong>samples<\/strong>, if each <strong>sample<\/strong> included 4 scores. Looks slightly better, but still not very convincing. With a <strong>sample<\/strong> size of 7, it looks a bit better. Once our <strong>sample<\/strong> size is 10, we at least have something pretty close. Mathematically speaking, as long as the <strong>sample<\/strong> size is no smaller than 30, then the assumption of normality holds. The other way we can reasonably make the normality assumption is if we know the population itself follows a normal curve. In that case, even if individual <strong>samples<\/strong> do not have a nice shaped histogram, that is okay, because the normality assumption is one apply to the <strong>population<\/strong> in question, not to the <strong>sample<\/strong> itself. <\/span>\r\n\r\n<span style=\"text-align: initial;font-size: 1em\">Now, you can play around with an <a href=\"https:\/\/onlinestatbook.com\/stat_sim\/sampling_dist\/\" target=\"_blank\" rel=\"noopener\">online demonstration<\/a> so you can really convince yourself that the <strong>central limit theorem<\/strong> works in practice. The goal here is to see what <strong>sample<\/strong> size is sufficient to generate a histogram that closely approximates a normal curve. And to trust that even if real-life data look wonky, the normal curve may still be a reasonable model for data analysis for purposes of inference.<\/span>\r\n<h4><a id=\"CLT_return\"><\/a><a href=\"#CLT\">Concept Practice: Central Limit Theorem<\/a><\/h4>\r\n<h1><a id=\"4b\"><\/a>4b. Hypothesis Testing<\/h1>\r\n<h6 style=\"text-align: right\"><a href=\"#video4a\">video lesson<\/a><\/h6>\r\nWe are finally ready for your first introduction to a formal decision making procedure often used in statistics, known as <strong>[pb_glossary id=\"259\"]hypothesis testing[\/pb_glossary]<\/strong>.\r\n\r\nIn this course, we started off with descriptive statistics, so that you would become familiar with ways to summarize the important characteristics of datasets. Then we explored the concepts standardizing scores, and relating those to probability as area under the normal curve model. With all those tools, we are now ready to make something!\r\n\r\n[caption id=\"attachment_268\" align=\"aligncenter\" width=\"1024\"]<img class=\"wp-image-268 size-large\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.11-1024x576.png\" alt=\"\" width=\"1024\" height=\"576\" \/> <em>\"(not my) toolbox\" by erix! is licensed under CC BY 2.0 \"Dovetail Dresser\" by Didriks is licensed under CC BY 2.0<\/em>[\/caption]\r\n\r\nOkay, not furniture, exactly, but decisions.\r\n\r\nWe are now into the portion of the course that deals with inferential statistics. Just to get you thinking in terms of making decisions on the basis of data, let us take a slightly silly example. Suppose I have discovered a pill that cures hangovers!\r\n\r\n<img class=\"alignright wp-image-263\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.7-768x1024.jpg\" alt=\"Tiberius\" width=\"447\" height=\"596\" \/>\r\n\r\nWell, it greatly lessened symptoms of hangover in 10 of the 15 people I tested it on. I am charging 50 dollars per pill. Will you buy it the next time you go out for a night of drinking? Or recommend it to a friend? \u2026 If you said yes, I wonder if you are thinking very critically? Should we think about the cost-benefit ratio here on the basis of what information you have? If you said no, I bet some of the doubts I bring up popped to your mind as well. If 10 out of 15 people saw lessened symptoms, that\u2019s 2\/3 of people \u2013 so some people saw no benefits. Also, what does \u201cgreatly lessened symptoms of hangover\u201d mean? Which symptoms? How much is greatly? Was the reduction by two or more standard deviations from the mean? Or was it less than one standard deviation improvement? Given the cost of 50 dollars per pill, I have to say I would be skeptical about buying it without seeing some statistics!\r\n\r\nOn this list is a preview of the basic concepts to which you will be introduced as we go through the rest of this chapter.\r\n<div class=\"textbox textbox--learning-objectives\"><header class=\"textbox__header\">\r\n<p class=\"textbox__title\">Hypothesis Testing Basic Concepts<\/p>\r\n\r\n<\/header>\r\n<div class=\"textbox__content\">\r\n<ul>\r\n \t<li>Hypothesis<\/li>\r\n \t<li>Null Hypothesis<\/li>\r\n \t<li>Research Hypothesis (alternative hypothesis)<\/li>\r\n \t<li>Statistical significance<\/li>\r\n \t<li>Conventional levels of significance<\/li>\r\n \t<li>Cutoff sample score (critical value)<\/li>\r\n \t<li>Directional vs. non-directional hypotheses<\/li>\r\n \t<li>One-tailed and two-tailed tests<\/li>\r\n \t<li>Type I and Type II errors<\/li>\r\n<\/ul>\r\n<\/div>\r\n<\/div>\r\nYou can see that there are lots of new concepts to master. In my experience, each concept makes the most sense in context, within its place in the <strong>hypothesis testing<\/strong> workflow. We will start with defining our <strong>null<\/strong> and <strong>research hypotheses<\/strong>, then discuss the <strong>[pb_glossary id=\"271\"]levels of statistical significance[\/pb_glossary]<\/strong> and their conventional usage. Next, we will look at how to find the <strong>[pb_glossary id=\"270\"]cutoff sample score[\/pb_glossary]<\/strong> that will form the critical value for our decision criterion. We will look at how that differs for <strong>directional<\/strong> vs. <strong>non-directional hypotheses<\/strong>, which will lend themselves to <strong>one-<\/strong> or <strong>two-tailed<\/strong> <strong>tests<\/strong>, respectively.\r\n\r\nThe <strong>hypothesis testing<\/strong> procedure, or workflow, can be broken down into five discrete steps.\r\n<div class=\"textbox textbox--exercises\"><header class=\"textbox__header\">\r\n<p class=\"textbox__title\">Steps of Hypothesis Testing<\/p>\r\n\r\n<\/header>\r\n<div class=\"textbox__content\">\r\n<ol>\r\n \t<li>Restate question as a research hypothesis and a null hypothesis about populations.<\/li>\r\n \t<li>Determine characteristics of the comparison distribution.<\/li>\r\n \t<li>Determine the cutoff sample score on the comparison distribution at which the null hypothesis should be rejected.<\/li>\r\n \t<li>Determine your sample\u2019s score on the comparison distribution.<\/li>\r\n \t<li>Decide whether to reject the null hypothesis.<\/li>\r\n<\/ol>\r\n<\/div>\r\n<\/div>\r\nThese steps are something we will be using pretty much the rest of the semester, so it is worth memorizing them now. My favourite approach to that is to create a mnemonic device. I recommend the following key words from which to form your mnemonic device: hypothesis, characteristics, cutoff, score, and decide. Not very memorable? Try association those with more memorable words that start with the same letter or sound. How about \u201c<em>Happy Chickens Cure Sad Days<\/em>.\u201d Or you can put the words into a <a href=\"https:\/\/www.mnemonicgenerator.com\/\" target=\"_blank\" rel=\"noopener\">mnemonic device generator<\/a> on the internet and get something truly bizarre. I just tried one and got \u201c<em>Hairless Carsick Chewbacca Slapped Demons<\/em>\u201d. Another good one: \u201c<em>Hamlet Chose Cranky Sushi Drunkenly<\/em>.\u201d Anyway, you play around with it or brainstorm until you hit upon one that works for you. Who knew statistics could be this much fun!\r\n\r\nThe first step in <strong>hypothesis testing<\/strong> is always to formulate hypotheses. The first rule that will help you do so correctly, is that hypotheses are always about <strong>populations<\/strong>. We study samples in order to make conclusions about populations, so our predictions should be about the populations themselves. First, we define <strong>population<\/strong> 1 and <strong>population<\/strong> 2. <strong>Population<\/strong> 1 is always defined as people like the ones in our research study, the ones we are truly interested in. <strong>Population<\/strong> 2 is the comparison <strong>population<\/strong>, the status quo to which we are looking to compare our research <strong>population<\/strong>. Now, remember, when referring to <strong>populations<\/strong>, we always use Greek letters. So if we formulate our hypotheses in symbols, we need to use Greek letters.\r\n\r\n<img class=\"wp-image-274 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.12-1024x163.png\" alt=\"\" width=\"524\" height=\"83\" \/>\r\n\r\nIt is a good idea to state our hypotheses both in symbols and in words. We need to make them specific and disprovable. If you follow my tips, you will have it down with just a little practice.\r\n\r\nWe need to state two hypotheses. First, we state the <strong>[pb_glossary id=\"275\"]research hypothesis[\/pb_glossary]<\/strong>, which is sometimes referred to as the alternative hypothesis. The <strong>research hypothesis<\/strong> (often called the alternative hypothesis) is a statement of inequality, or that Something happened! This hypothesis makes the prediction that the <strong>population<\/strong> from which the research sample came is different from the comparison <strong>population<\/strong>. In other words, there is a really high <strong>probability<\/strong> that the sample comes from a different distribution than the comparison one.\r\n\r\nThe <strong>[pb_glossary id=\"277\"]null hypothesis[\/pb_glossary]<\/strong>, on the other hand, is a statement of equality, or that nothing happened. This hypothesis makes the prediction that the <strong>population<\/strong> from which sample came is not different from the comparison <strong>population<\/strong>. We set up the <strong>null hypothesis<\/strong> as a so-called straw man, that we hope to tear down. Just remember, null means nothing \u2013 that nothing is different between the <strong>populations<\/strong>.\r\n\r\n&nbsp;\r\n\r\nStep two of hypothesis testing is to determine the characteristics of the comparison distribution. This is where our descriptive statistics, the mean and standard deviation, come in. We need to ensure our normal curve model to which we are comparing our research <strong>sample<\/strong> is mapped out according to the particular characteristics of the <strong>population<\/strong> of comparison, which is <strong>population<\/strong> 2.\r\n\r\nNext it is time to set our decision rule. Step 3 is to determine the <strong>cutoff sample score<\/strong>, which is derived from two pieces of information. The first is the conventional <strong>significance level<\/strong> that applies. By convention, the <strong>probability<\/strong> level that we are willing to accept as a risk that the score from our research sample might occur by random chance within the comparison distribution is set to one of three levels: 10%, 5%, or 1%. The most common choice of <strong>significance level<\/strong> is 5%. Typically the <strong>significance level<\/strong> will be provided to you in the problem for your statistics courses, but if it is not, just default to a <strong>significance level<\/strong> of .05. Sometimes researchers will choose a more conservative <strong>significance level<\/strong>, like 1%, if they are particularly risk averse. If the researcher chooses a 10% <strong>significance level<\/strong>, they are likely conducting a more exploratory study, perhaps a pilot study, and are not too worried about the <strong>probability<\/strong> that the score might be fairly common under the comparison distribution.\r\n\r\nThe second piece of information we need to know in order to find our <strong>cutoff sample score<\/strong> is which tail we are looking at. Is this a <strong>[pb_glossary id=\"284\"]directional hypothesis[\/pb_glossary]<\/strong>, and thus <strong>[pb_glossary id=\"280\"]one-tailed test[\/pb_glossary]<\/strong>? Or a <strong>[pb_glossary id=\"285\"]non-directional hypothesis[\/pb_glossary]<\/strong>, and thus a <strong>[pb_glossary id=\"281\"]two-tailed test[\/pb_glossary]<\/strong>? This depends on the <strong>research hypothesis<\/strong> from step 1. Look for directional keywords in the problem. If the researcher prediction involves words like \u201cgreater than\u201d or \u201clarger than\u201d, this signals that we should be doing a <strong>one-tailed test<\/strong> and that our <strong>cutoff<\/strong> <strong>sample score <\/strong>should be in the top tail of the distribution. If the researcher prediction involves words like \u201clower than\u201d or \u201csmaller than\u201d, this signals that we should be doing a <strong>one-tailed test<\/strong> and that our <strong>cutoff<\/strong> <strong>sample score <\/strong>should be in the bottom tail of the distribution. If the prediction is neutral in directionality, and uses a word like \u201cdifferent\u201d, that signals a <strong>non-directional hypothesis<\/strong>. In that case, we would need to use a two-tailed test, and our cutoff scores would need to be indicated on both tails of the distribution. To do that, we take our area under the curve, which matches our <strong>significance level<\/strong>, and split it into both tails.\r\n\r\n<img class=\"aligncenter wp-image-265 \" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.8-1-1024x558.jpg\" alt=\"\" width=\"554\" height=\"302\" \/>\r\n\r\nFor example, if we have a <strong>two-tailed test<\/strong> with a .05 <strong>significance level<\/strong>, then we would split the 5% area under the curve into the two tails, so two and a half percent in each tail.\r\n<h4><a id=\"tails_return\"><\/a><a href=\"#tails\">Concept Practice: deciding on one-tailed vs. two-tailed tests<\/a><\/h4>\r\nWe can find the Z-score that forms the border of the tail area we have identified based on <strong>significance level<\/strong> and directionality by looking it up in a <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/back-matter\/normal-curve-z-area-tables\/\" target=\"_blank\" rel=\"noopener\">table<\/a> or an <a href=\"https:\/\/statpages.info\/pdfs.html\" target=\"_blank\" rel=\"noopener\">online calculator<\/a>. I always recommend mapping this <strong>cutoff score<\/strong> onto a drawing of the comparison distribution as shown above. This should help you visualize the setup of the <strong>hypothesis test<\/strong> clearly and accurately.\r\n<h4><a id=\"hypTesting_return\"><\/a><a href=\"#hypTesting\">Concept Practice: inference through hypothesis testing<\/a><\/h4>\r\nThe next step in the <strong>hypothesis testing<\/strong> procedure is to determine your <strong>sample\u2019s<\/strong> score on the comparison distribution. To do this, we calculate a test statistic from the <strong>sample<\/strong> raw score, mark it on the comparison distribution, and determine whether it falls in the shaded tail or not. In reality, we would always have a sample with more than one score in it. However, for the sake of keeping our test statistic formula a familiar one, we will use a sample size of one. We will use our Z-score formula to translate the <strong>sample\u2019s<\/strong> raw score into a Z-score \u2013 in other words, we will figure out how many standard deviations above or below the comparison distribution's mean the <strong>sample<\/strong> score is.\r\n\r\n[latex] \\[Z=\\frac{X-M}{SD}\\] [\/latex]\r\n\r\nFinally, it\u2019s time to decide whether to <strong>[pb_glossary id=\"288\"]reject the null hypothesis[\/pb_glossary]<\/strong>. This decision is based on whether our sample\u2019s data point was more extreme than the <strong>cutoff score<\/strong>, in other words, \u201cdid it fall in the shaded tail?\u201d If the <strong>sample<\/strong> score is more extreme than the <strong>cutoff score<\/strong>, then we must reject the null hypothesis. Our research hypothesis is supported! (Not proven\u2026 remember, there is still some probability that that score could have occurred randomly within the comparison distribution.) But it is sound to say that it appears quite likely that the population from which our sample came is different from the comparison population. Another way to express this decision is to say that the result was <strong>[pb_glossary id=\"291\"]statistically significant[\/pb_glossary]<\/strong>, which is to say that there is less than a 5% chance of this result occurring randomly within the comparison distribution (here I just filled in the blank with the significance level).\r\n\r\nWhat if the research sample score did not fall in the shaded tail? In the case that the sample score is less extreme than the <strong>cutoff score<\/strong>, then our <strong>research hypothesis<\/strong> is not supported. We <strong>[pb_glossary id=\"289\"]do not reject the null hypothesis[\/pb_glossary]<\/strong>. It appears that the <strong>population<\/strong> from which our <strong>sample<\/strong> came is not different from the comparison <strong>population<\/strong>. Note that we do not typically express this result as \u201caccept the null hypothesis\u201d or \u201cwe have proved the null hypothesis\u201d. From this test, we do not have evidence that the <strong>null hypothesis<\/strong> is correct, rather we simply did not have enough evidence to reject it. Another way to express this decision is to say that the result was not <strong>statistically significant<\/strong>, which is to say that there is more than a 5% chance of this result occurring randomly within the comparison distribution (here I just used the most common <strong>significance level<\/strong>).\r\n<h4><a id=\"interpretingHyp_return\"><\/a><a href=\"#interpretingHyp\">Concept Practice: interpreting conclusions of hypothesis tests<\/a><\/h4>\r\nSo we have described the <strong>hypothesis testing<\/strong> process from beginning to end. The whole process of null <strong>hypothesis testing<\/strong> can feel like pretty tortured logic at first. So let us zoom out, and look at the whole process another way. Essentially what we are seeking to do in such a <strong>hypothesis test<\/strong> is to compare two <strong>populations<\/strong>. We want to find out if the <strong>populations<\/strong> are distinct enough to confidently state that there is a difference between <strong>population<\/strong> 1 and <strong>population<\/strong> 2. In our example, we wanted to know if the <strong>population<\/strong> of people using a new medication, <strong>population<\/strong> 1, sleep longer than the <strong>population<\/strong> of people who are not using that new medication, <strong>population<\/strong> 2. We ended up finding that the research evidence to hand suggests <strong>population<\/strong> 1\u2019s distribution is distinct enough from <strong>population<\/strong> 2 that we could <strong>reject the null hypothesis<\/strong> of similarity.\r\n\r\n<img class=\"wp-image-266 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.9-1-1024x411.jpg\" alt=\"\" width=\"546\" height=\"219\" \/>\r\n\r\nIn other words, we were able to conclude that the difference between the centres of the two distributions was <strong>statistically significant<\/strong>.\r\n\r\nIf, on the other hand, the distributions were a bit less distinct, we would not have been able to make that claim of a significant difference.\r\n\r\n<img class=\"wp-image-267 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.10-1-1024x605.jpg\" alt=\"\" width=\"377\" height=\"223\" \/>\r\n\r\nWe would not have <strong>rejected the null hypothesis<\/strong> if evidence indicated the <strong>populations<\/strong> were too similar.\r\n\r\nJust how different do the two distributions need to be? That criterion is set by the <strong>cutoff score<\/strong>, which depends on the <strong>significance level<\/strong>, and whether it is a <strong>one-tailed<\/strong> or <strong>two-tailed<\/strong> <strong>hypothesis test<\/strong>.\r\n<h4><a id=\"Hyptest_return\"><\/a><a href=\"#Hyptest\">Concept Practice: Putting hypothesis test elements together<\/a><\/h4>\r\nThat was a lot of new concepts to take on! As a reward, assuming you enjoy memes, there are a plethora of <a href=\"https:\/\/www.google.com\/search?q=statistics+memes&amp;rlz=1C1CHBF_enCA784CA784&amp;oq=statistics+memes&amp;aqs=chrome..69i57j0i433i512j0i512l3j69i60l3.2945j1j7&amp;sourceid=chrome&amp;ie=UTF-8\" target=\"_blank\" rel=\"noopener\">statistics memes<\/a>, some of which you may find funny now that you have made it into inferential statistics territory. Welcome to the exclusive club of people who have this rather peculiar sense of humour. Enjoy!\r\n<h1>Chapter Summary<\/h1>\r\nIn this chapter we examined <strong>probability<\/strong> and how it can be used to make inferences about data in the framework of <strong>hypothesis testing<\/strong>. We now have a sense of how two <strong>populations<\/strong> can be compared and the difference between their means evaluated for <strong>statistical significance<\/strong>.\r\n\r\nKey Terms:\r\n<table class=\"no-lines\" style=\"border-collapse: collapse;width: 100%;height: 76px\" border=\"0\">\r\n<tbody>\r\n<tr style=\"height: 15px\">\r\n<td style=\"width: 27.141%;height: 15px\"><strong>probability<\/strong><\/td>\r\n<td style=\"width: 29.5125%;height: 15px\"><strong>research hypothesis\u00a0<\/strong><\/td>\r\n<td style=\"width: 43.8734%;height: 15px\"><strong>one-tailed test<\/strong><\/td>\r\n<\/tr>\r\n<tr style=\"height: 15px\">\r\n<td style=\"width: 27.141%;height: 15px\"><strong>central limit theorem<\/strong><\/td>\r\n<td style=\"width: 29.5125%;height: 15px\"><strong>null hypothesis <\/strong><\/td>\r\n<td style=\"width: 43.8734%;height: 15px\"><strong>non-directional hypothesis<\/strong><\/td>\r\n<\/tr>\r\n<tr style=\"height: 14px\">\r\n<td style=\"width: 27.141%;height: 16px\"><strong>population<\/strong><\/td>\r\n<td style=\"width: 29.5125%;height: 16px\"><strong>cutoff sample score<\/strong><\/td>\r\n<td style=\"width: 43.8734%;height: 16px\"><strong>statistical significance<\/strong><\/td>\r\n<\/tr>\r\n<tr style=\"height: 15px\">\r\n<td style=\"width: 27.141%;height: 15px\"><strong>sample<\/strong><\/td>\r\n<td style=\"width: 29.5125%;height: 15px\"><strong>significance level<\/strong><\/td>\r\n<td style=\"width: 43.8734%;height: 15px\"><strong>reject the null hypothesis<\/strong><\/td>\r\n<\/tr>\r\n<tr style=\"height: 15px\">\r\n<td style=\"width: 27.141%;height: 15px\"><strong>hypothesis testing<\/strong><\/td>\r\n<td style=\"width: 29.5125%;height: 15px\"><strong>directional hypothesis<\/strong><\/td>\r\n<td style=\"width: 43.8734%;height: 15px\"><strong>do not<\/strong> <strong>reject the null hypothesis<\/strong><\/td>\r\n<\/tr>\r\n<\/tbody>\r\n<\/table>\r\n<h1>Concept Practice<\/h1>\r\n<a id=\"defineProb\"><\/a>[h5p id=\"17\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#defineProb_return\">text<\/a><\/h6>\r\n<a id=\"calcProb1\"><\/a>[h5p id=\"13\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#calcProb1_return\">text<\/a><\/h6>\r\n<a id=\"calcProb2\"><\/a>[h5p id=\"16\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#calcProb2_return\">text<\/a><\/h6>\r\n<a id=\"luckystreak\"><\/a>[h5p id=\"15\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#lucky_return\">text<\/a><\/h6>\r\n<a id=\"areaProb\"><\/a>[h5p id=\"18\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#areaProb_return\">text<\/a><\/h6>\r\n<a id=\"findPercentile\"><\/a>[h5p id=\"19\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#findPercentile_return\">text<\/a><\/h6>\r\n<a id=\"interpretPercentile\"><\/a>[h5p id=\"21\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#interpretPercentile_return\">text<\/a><\/h6>\r\n<a id=\"applyProb\"><\/a>[h5p id=\"20\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#applyProb_return\">text<\/a><\/h6>\r\n<a id=\"statsProb\"><\/a>[h5p id=\"11\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#statsProb_return\">text<\/a><\/h6>\r\n<a id=\"CLT\"><\/a>[h5p id=\"36\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#CLT_return\">text<\/a><\/h6>\r\n<a id=\"tails\"><\/a>[h5p id=\"26\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#tails_return\">text<\/a><\/h6>\r\n<a id=\"hypTesting\"><\/a>[h5p id=\"22\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#hypTesting_return\">text<\/a><\/h6>\r\n<a id=\"interpretingHyp\"><\/a>[h5p id=\"23\"]\r\n\r\n[h5p id=\"24\"]\r\n\r\n[h5p id=\"25\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#interpretingHyp_return\">text<\/a><\/h6>\r\n<a id=\"Hyptest\"><\/a>[h5p id=\"29\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#Hyptest_return\">text<\/a><\/h6>\r\nReturn to <a href=\"#4a\">4a. Probability and Inferential Statistics<\/a>\r\n<h6>Try interactive <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/chapter\/worksheet-4a\/\" target=\"_blank\" rel=\"noopener\">Worksheet 4a<\/a> or download <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/01\/4a.-Class-worksheet.pdf\" target=\"_blank\" rel=\"noopener\">Worksheet 4a<\/a><\/h6>\r\nReturn to <a href=\"#4b\">4b. Hypothesis Testing<\/a>\r\n<h6>Try interactive <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/chapter\/worksheet-4b\/\" target=\"_blank\" rel=\"noopener\">Worksheet 4b<\/a>\u00a0or download <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2023\/12\/4b.-Class-worksheet.pdf\">Worksheet 4b<\/a><\/h6>\r\n<a id=\"video4a\"><\/a>video 4a1\r\n\r\n[video width=\"1440\" height=\"1080\" mp4=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/08\/Chapter-4a1.mp4\"][\/video]\r\n\r\nvideo 4a2\r\n\r\n[video width=\"1440\" height=\"1080\" mp4=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Chapter-4a2.mp4\"][\/video]\r\n\r\n<a id=\"video4b\"><\/a>video 4b1\r\n\r\n[video width=\"1440\" height=\"1080\" mp4=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Chapter-4b1.mp4\"][\/video]\r\n\r\nvideo 4b2\r\n\r\n[video width=\"1440\" height=\"1080\" mp4=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/08\/Chapter-4b2.mp4\"][\/video]","rendered":"<h1><a id=\"4a\"><\/a>4a. Probability and Inferential Statistics<\/h1>\n<h6 style=\"text-align: right\"><a href=\"#video4a\">video lesson<\/a><\/h6>\n<p>In this chapter, we will focus on connecting concepts of <a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_233\"><strong>probability<\/strong> <\/a> with the logic of inferential statistics.<br \/>\n<span class=\"pullquote-right\">&#8220;The whole problem with the world is that fools and fanatics are always so certain of themselves, and wiser people so full of doubts.&#8221;<br \/>\n\u2014 Bertrand Russel (1872-1970)<\/span><\/p>\n<p>These notable quotes represent why <strong>probability<\/strong> is critical for a basic understanding of scientific reasoning.<\/p>\n<p><span class=\"pullquote-left\">&#8220;Medicine is a science of uncertainty and an art of probability.&#8221;<br \/>\n\u2014 William Osler (1849\u20131919)<\/span>In many ways, the process of postsecondary education is all about instilling a sense of doubt and wonder, and the ability to estimate <strong>probabilities<\/strong>. As a matter of fact, that essentially sums up the entire reason why you are in this course. So let us tackle <strong>probability<\/strong>.<\/p>\n<p>We will be keeping our coverage of <strong>probability<\/strong> to a very simple level, because the introductory statistics we will cover rely on only simple <strong>probability<\/strong>. That said, I encourage you to read further on compound and conditional <strong>probabilities<\/strong>, because they will certainly make you smarter at real-life decision making. We will briefly touch on examples of how bad people can be at using <strong>probability<\/strong> in real life, and we will then address what probability has to do with inferential statistics. Finally, I will introduce you to the <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_235\">central limit theorem<\/a><\/strong>. This is probably one of the heftiest math concepts in the course, but worry not. Its implications are easy to learn, and the concepts behind it can be demonstrated empirically in the interactive exercises.<\/p>\n<p>First, we need to define <strong>probability<\/strong>. In a situation where several different outcomes are possible, the probability of any specific outcome is a fraction or proportion of all possible outcomes. Another way of saying that is this. If you wish to answer the question, \u201cWhat are the chances that outcome would have happened?\u201d, you can calculate the <strong>probability<\/strong> as the ratio of possible successful outcomes to all possible outcomes.<\/p>\n<h4><a id=\"defineProb_return\"><\/a><a href=\"#defineProb\">Concept Practice: define probability<\/a><\/h4>\n<p>People often use the rolling of dice as examples of simple <strong>probability<\/strong> problems.<\/p>\n<figure style=\"width: 371px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"\" src=\"https:\/\/live.staticflickr.com\/62\/201651074_30058a36ac_b.jpg\" alt=\"Dice\" width=\"371\" height=\"278\" \/><figcaption class=\"wp-caption-text\">&#8220;Dice&#8221; by matsuyuki is licensed under CC BY-SA 2.0<\/figcaption><\/figure>\n<p>If you were to roll one typical die, which has a number on each side from 1 to 6, then the simple probability of rolling a 1 would be 1\/6. There are six possible outcomes, but only 1 of them is the successful outcome, that of rolling a 1.<\/p>\n<h4><a id=\"calcProb1_return\"><\/a><a href=\"#calcProb1\">Concept Practice: calculate probability<\/a><\/h4>\n<p>Another common example used to introduce simple <strong>probability<\/strong> is cards. In a standard deck of casino cards, there are 52 cards. There are 4 aces in such a deck of cards (Aces are the \u201c1\u201d card, and there is 1 in each suit \u2013 hearts, spades, diamonds and clubs.)<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-238\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.2-1024x576.png\" alt=\"\" width=\"539\" height=\"303\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.2-1024x576.png 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.2-300x169.png 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.2-768x432.png 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.2-65x37.png 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.2-225x127.png 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.2-350x197.png 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.2.png 1280w\" sizes=\"auto, (max-width: 539px) 100vw, 539px\" \/><\/p>\n<p>If you were to ask the question \u201cwhat is the <strong>probability<\/strong> that a card drawn at random from a deck of cards will be an ace?\u201d, and you know all outcomes are equally likely, the <strong>probability<\/strong> would be the ratio of the number of times one could draw and ace divided by the number of all possible outcomes. In this example, then, the <strong>probability<\/strong> would be 4\/52. This ratio can be converted into a decimal: 4 divided by 52 is 0.077, or 7.7%. (Remember, to turn a decimal to a percent, you need to move the decimal place twice to the right.)<\/p>\n<h4><a id=\"calcProb2_return\"><\/a><a href=\"#calcProb2\">Concept Practice: calculate probability<\/a><\/h4>\n<p><strong>Probability<\/strong> seems pretty straightforward, right? But people often misunderstand <strong>probability<\/strong> in real life. Take the idea of the lucky streak, for example. Let\u2019s say someone is rolling dice and they get 4 6\u2019s in a row. Lots of people might say that\u2019s a lucky streak and they might go as far as to say they should continue, because their luck is so good at the moment! According to the rules of <strong>probability<\/strong>, though, the next die roll has a 1\/6 chance of being a 6, just like all the others. True, the <strong>probability<\/strong> of a 4-in-a-row streak occurring is fairly slim: 1\/6 x 1\/6 x 1\/6 x 1\/6. But the fact is that this rare event does not predict future events (unless it is an unfair die!). Each time you roll a die, the <strong>probability<\/strong> of that event remains the same. That is what the human brain seems to have a really hard time accepting.<\/p>\n<h4><a id=\"lucky_return\"><\/a><a href=\"#luckystreak\">Concept Practice: lucky streak<\/a><\/h4>\n<p>When someone makes a prediction attached to a certain probability (e.g. there is only a 1% chance of an earthquake in the next week), and then that event occurs in spite of that low <strong>probability<\/strong> estimate (e.g. there is actually an earthquake the day after the prediction was made)\u2026 was that person wrong? No, not really, because they allowed for the possibility. Had they said there was a 0% chance, they would have been wrong.<\/p>\n<p><strong>Probabilities<\/strong> are often used to express likelihood of outcomes under conditions of uncertainty. Like Bertrand Russell said, wise people rarely speak in terms of certainties. Because people so often misunderstand <strong>probability<\/strong>, or find irrational actions so hard to resist despite some understanding of <strong>probability<\/strong>, decision making in the realm of sciences needs to be designed to combat our natural human tendencies. What we are discussing now in terms of how to think about and calculate <strong>probabilities<\/strong> will form a core component of our decision-making framework as we move forward in the course.<\/p>\n<p>Now, let\u2019s take a look at how <strong>probability<\/strong> is used in statistics.<\/p>\n<h4><a id=\"areaProb_return\"><\/a><a href=\"#areaProb\">Concept Practice: area under normal curve as probability<\/a><\/h4>\n<p>We saw that percentiles are expressions of area under a normal curve. Areas under the curve can be expressed as <strong>probability<\/strong>, too. For example, if we say the 50th percentile for IQ is 100, that can be expressed as: &#8220;If I chose a person at random, there is a 50% chance that they will have an IQ score below 100.\u201d<\/p>\n<p>&nbsp;<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-246\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.3-1024x553.jpg\" alt=\"\" width=\"534\" height=\"289\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.3-1024x553.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.3-300x162.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.3-768x415.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.3-1536x829.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.3-65x35.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.3-225x121.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.3-350x189.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.3.jpg 1745w\" sizes=\"auto, (max-width: 534px) 100vw, 534px\" \/><\/p>\n<p>If we find the 84th percentile for IQ is 115 there is another way to say that \u201cIf I chose a person at random, there is an 84% chance that they will have an IQ score below 115.\u201d<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-245\" style=\"font-size: 1em\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.4-1024x594.jpg\" alt=\"\" width=\"521\" height=\"302\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.4-1024x594.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.4-300x174.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.4-768x445.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.4-1536x890.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.4-65x38.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.4-225x130.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.4-350x203.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.4.jpg 1696w\" sizes=\"auto, (max-width: 521px) 100vw, 521px\" \/><\/p>\n<h4><a id=\"findPercentile_return\"><\/a><a href=\"#findPercentile\">Concept Practice: find percentiles<\/a><\/h4>\n<p>Any time you are dealing with area under the normal curve, I encourage you to express that percentage in terms of <strong>probabilities<\/strong>. That will help you think clearly about what that area under the curve means once we get into the exercise of making decisions based on that information.<\/p>\n<h4><a id=\"interpretPercentile_return\"><\/a><a href=\"#interpretPercentile\">Concept Practice: interpreting percentile as probability<\/a><\/h4>\n<p><strong>Probabilities<\/strong>, of course, range from 0 to 1 as proportions or fractions, and from 0% to 100% when expressed in percentage terms. In inferential statistics, we often express in terms of <strong>probability<\/strong> the likelihood that we would observe a particular score under a given normal curve model.<\/p>\n<h4><a id=\"applyProb_return\"><\/a><a href=\"#applyProb\">Concept Practice: applying probability<\/a><\/h4>\n<p>Although I encourage you to think of <strong>probabilities<\/strong> as percentages, the convention in statistics is to report to the probability of a score as a proportion, or decimal. The symbol used for \u201cprobability of score\u201d is <em>p<\/em>. In statistics, the interpretation of \u201c<em>p<\/em>\u201d is a delicate subject. Generations of researchers have been lazy in our understanding of what \u201c<em>p<\/em>\u201d: tells us, and we have tended to over-interpret this statistic. As we begin to work with \u201c<em>p<\/em>\u201d, I will ask you to memorize a mantra that will help you report its meaning accurately. For now, just keep in mind that most psychologists and psychology students <em>still<\/em> make mistakes in how they express and understand the meaning of \u201c<em>p<\/em>\u201d values. This will take time and effort to fix, but I am confident that your generation will learn to do better at a precise and careful understanding of what statistics like \u201c<em>p<\/em>\u201d tell us\u2026 and what they do not.<\/p>\n<p>To give you a sense of what a statement of <em>p<\/em> &lt; .05 might mean, let us think back to our rat weights example.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-201 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.10-1024x624.jpg\" alt=\"\" width=\"596\" height=\"363\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.10-1024x624.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.10-300x183.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.10-768x468.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.10-1536x936.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.10-65x40.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.10-225x137.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.10-350x213.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.10.jpg 1908w\" sizes=\"auto, (max-width: 596px) 100vw, 596px\" \/><\/p>\n<p>If I were to take a rat from our high-grain food group and place it on the distribution of untreated rat weights, and if it placed at <em>Z<\/em> = .9, we could look at the area under the curve from that point and above. That would tell us how likely it would be to observe such a heavy rat in the general population of nontreated rats &#8212; those that eat a normal diet.<\/p>\n<p>Think of it this way. When we select a rat from our treatment group (those that ate the grain-heavy diet), and it is heavier than the average for a nontreated rat, there are two possible explanations for that observation. One is that the diet made him that way. As a scientist whose hypothesis is that a grain-heavy diet will make the rats weigh more, I\u2019m actually motivated to interpret the observation that way. I want to believe this event is meaningful, because it is consistent with my hypothesis! But the other possibility is that, by random chance, we picked a rat that was heavy to begin with. There are plenty of rats in the distribution of nontreated rats that were at least that heavy. So there is always some <strong>probability<\/strong> that we just randomly selected a heavier rat. In this case, if my treated rat\u2019s weight was less than one standard deviation above the mean, we saw in the chapter on normal curves that the <strong>probability<\/strong> of observing a rat weight that high or higher in the nontreated population was about 18%. That is not so unusual. It would not be terribly surprising if that outcome were simply the result of random chance rather than a result of the diet the rat had been eating.<\/p>\n<p>If, on the other hand, the rat we measured was 2.5 standard deviations above the mean, the tail <strong>probability<\/strong> beyond that <strong>Z-score<\/strong> would be vanishingly small.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-202\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.11-1024x645.jpg\" alt=\"\" width=\"605\" height=\"381\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.11-1024x645.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.11-300x189.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.11-768x484.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.11-1536x967.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.11-65x41.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.11-225x142.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.11-350x220.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/07\/Fig-3.11.jpg 1882w\" sizes=\"auto, (max-width: 605px) 100vw, 605px\" \/><\/p>\n<p>The <strong>probability<\/strong> of observing such a rat weight in the nontreated population is very low, so it is far less likely that observation can be accounted for just by random chance alone. As we accumulate more evidence, the <strong>probability<\/strong> they could have come at random from the nontreated <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_248\">population<\/a><\/strong> will weigh into our decision making about whether the grain-heavy diet indeed causes rats to become heavier. This is the way <strong>probabilities<\/strong> are used in the process of <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_259\">hypothesis testing<\/a><\/strong>, the logic of inferential statistics that we will look at soon.<\/p>\n<h4><a id=\"statsProb_return\"><\/a><a href=\"#statsProb\">Concept Practice: statistics as probability<\/a><\/h4>\n<p>Now that you have seen the relevance of <strong>probability<\/strong> to the decision making process that comprises inferential statistics, we have one more major learning objective: to become familiar with the <strong>central limit theorem<\/strong>.<\/p>\n<p>However, before we get to the <strong>central limit theorem<\/strong>, we need to be clear on the distinction between two concepts:\u00a0 <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_249\">sample<\/a><\/strong> and <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_248\">population<\/a><\/strong>. In the world of statistics, the <strong>population<\/strong> is defined as all possible individuals or scores about which we would ideally draw conclusions. When we refer to the characteristics, or parameters, that describe a <strong>population<\/strong>, we will use Greek letters.\u00a0A <strong>sample<\/strong> is defined as the individuals or scores about which we are actually drawing conclusions. When we refer to the characteristics, or statistics, that describe a <strong>sample<\/strong>, we will use English letters.<\/p>\n<p>It is important to understand the difference between a <strong>population<\/strong> and a <strong>sample<\/strong>, and how they relate to one another, in order to comprehend the <strong>central limit theorem<\/strong> and its usefulness for statistics. From a <strong>population<\/strong> we can draw multiple <strong>samples<\/strong>. The larger <strong>sample<\/strong>, the more closely our <strong>sample<\/strong> will represent the <strong>population<\/strong>.<\/p>\n<p>Think of a Venn diagram.\u00a0<span style=\"text-align: initial;font-size: 1em\">There is a circle that is a <\/span><strong style=\"text-align: initial;font-size: 1em\">population<\/strong><span style=\"text-align: initial;font-size: 1em\">. Inside that large circle, you can draw an infinite number of smaller circles, each of which represents a <\/span><strong style=\"text-align: initial;font-size: 1em\">sample<\/strong><span style=\"text-align: initial;font-size: 1em\">.<\/span><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"size-medium wp-image-254 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.6-300x285.jpg\" alt=\"\" width=\"300\" height=\"285\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.6-300x285.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.6-1024x972.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.6-768x729.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.6-65x62.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.6-225x214.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.6-350x332.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.6.jpg 1129w\" sizes=\"auto, (max-width: 300px) 100vw, 300px\" \/><\/p>\n<p>The larger that inner circle, the more of the <strong>population<\/strong> it contains, and thus the more representative it is.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"size-medium wp-image-253 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.5-300x283.jpg\" alt=\"\" width=\"300\" height=\"283\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.5-300x283.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.5-1024x967.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.5-768x725.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.5-65x61.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.5-225x212.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.5-350x331.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.5.jpg 1096w\" sizes=\"auto, (max-width: 300px) 100vw, 300px\" \/><\/p>\n<p>Let us take a concrete example. A <strong>population<\/strong> might be the depression screening scores for all current postsecondary students in Canada. A sample from that <strong>population<\/strong> might be depression screening scores for 500 randomly selected postsecondary students from several institutions across Canada. That seems a more reasonable proportion of the two million students in the <strong>population<\/strong> than a <strong>sample<\/strong> that contains only 5 students. The 500 student <strong>sample<\/strong> has a better shot at adequately representing the entire <strong>population<\/strong> than does the 5 student <strong>sample<\/strong>, right? You can see that intuitively\u2026 and once you learn the <strong>central limit theorem<\/strong>, you will see the mathematical demonstration of the importance of <strong>sample<\/strong> size for representing the <strong>population<\/strong>.<\/p>\n<p>To conduct the inferential statistics we are using in this course, we will be using the normal curve model to estimate <strong>probabilities<\/strong> associated with particular scores. To do that, we need to assume that data are normally distributed. However, in real life, our data are almost never actually a perfect match for the normal curve.<\/p>\n<p>So how can we reasonably make the normality assumption? Here\u2019s the thing. The <strong>central limit theorem<\/strong> is a mathematical principle that assures us that the normality assumption is a reasonable one as long as we have a decent <strong>sample<\/strong> size.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignright wp-image-243\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.1-353x1024.jpg\" alt=\"\" width=\"298\" height=\"864\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.1-353x1024.jpg 353w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.1-103x300.jpg 103w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.1-529x1536.jpg 529w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.1-65x189.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.1-225x653.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.1-350x1016.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.1.jpg 620w\" sizes=\"auto, (max-width: 298px) 100vw, 298px\" \/><\/p>\n<p>According to the theorem, as long as we take a decent-sized <strong>sample<\/strong>, if we took many <strong>samples<\/strong> (10,000) of large enough size (30+) and took the mean each time, the distribution of those means w<span style=\"text-align: initial;font-size: 1em\">ill approach a normal distribution, even if the scores from each <\/span><strong style=\"text-align: initial;font-size: 1em\">sample<\/strong><span style=\"text-align: initial;font-size: 1em\"> are not normally distributed. To see this for yourself, take a look at the histograms shown on the right. The top histogram came from taking from a <\/span><strong style=\"text-align: initial;font-size: 1em\">population<\/strong><span style=\"text-align: initial;font-size: 1em\"> 10,000 <\/span><strong style=\"text-align: initial;font-size: 1em\">samples <\/strong><span style=\"text-align: initial;font-size: 1em\">of just one score each, and plotting them on a histogram. See how it has a flat, or rectangular shape? No way we could call that a shape approximating a normal<\/span><span style=\"text-align: initial;font-size: 1em\">\u00a0curve. Next is a histogram that came from taking the means of 10,000 <strong>samples<\/strong>, if each <strong>sample<\/strong> included 4 scores. Looks slightly better, but still not very convincing. With a <strong>sample<\/strong> size of 7, it looks a bit better. Once our <strong>sample<\/strong> size is 10, we at least have something pretty close. Mathematically speaking, as long as the <strong>sample<\/strong> size is no smaller than 30, then the assumption of normality holds. The other way we can reasonably make the normality assumption is if we know the population itself follows a normal curve. In that case, even if individual <strong>samples<\/strong> do not have a nice shaped histogram, that is okay, because the normality assumption is one apply to the <strong>population<\/strong> in question, not to the <strong>sample<\/strong> itself. <\/span><\/p>\n<p><span style=\"text-align: initial;font-size: 1em\">Now, you can play around with an <a href=\"https:\/\/onlinestatbook.com\/stat_sim\/sampling_dist\/\" target=\"_blank\" rel=\"noopener\">online demonstration<\/a> so you can really convince yourself that the <strong>central limit theorem<\/strong> works in practice. The goal here is to see what <strong>sample<\/strong> size is sufficient to generate a histogram that closely approximates a normal curve. And to trust that even if real-life data look wonky, the normal curve may still be a reasonable model for data analysis for purposes of inference.<\/span><\/p>\n<h4><a id=\"CLT_return\"><\/a><a href=\"#CLT\">Concept Practice: Central Limit Theorem<\/a><\/h4>\n<h1><a id=\"4b\"><\/a>4b. Hypothesis Testing<\/h1>\n<h6 style=\"text-align: right\"><a href=\"#video4a\">video lesson<\/a><\/h6>\n<p>We are finally ready for your first introduction to a formal decision making procedure often used in statistics, known as <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_259\">hypothesis testing<\/a><\/strong>.<\/p>\n<p>In this course, we started off with descriptive statistics, so that you would become familiar with ways to summarize the important characteristics of datasets. Then we explored the concepts standardizing scores, and relating those to probability as area under the normal curve model. With all those tools, we are now ready to make something!<\/p>\n<figure id=\"attachment_268\" aria-describedby=\"caption-attachment-268\" style=\"width: 1024px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-268 size-large\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.11-1024x576.png\" alt=\"\" width=\"1024\" height=\"576\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.11-1024x576.png 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.11-300x169.png 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.11-768x432.png 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.11-65x37.png 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.11-225x127.png 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.11-350x197.png 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.11.png 1280w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><figcaption id=\"caption-attachment-268\" class=\"wp-caption-text\"><em>&#8220;(not my) toolbox&#8221; by erix! is licensed under CC BY 2.0 &#8220;Dovetail Dresser&#8221; by Didriks is licensed under CC BY 2.0<\/em><\/figcaption><\/figure>\n<p>Okay, not furniture, exactly, but decisions.<\/p>\n<p>We are now into the portion of the course that deals with inferential statistics. Just to get you thinking in terms of making decisions on the basis of data, let us take a slightly silly example. Suppose I have discovered a pill that cures hangovers!<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignright wp-image-263\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.7-768x1024.jpg\" alt=\"Tiberius\" width=\"447\" height=\"596\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.7-768x1024.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.7-225x300.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.7-1152x1536.jpg 1152w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.7-1536x2048.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.7-65x87.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.7-350x467.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.7-scaled.jpg 1920w\" sizes=\"auto, (max-width: 447px) 100vw, 447px\" \/><\/p>\n<p>Well, it greatly lessened symptoms of hangover in 10 of the 15 people I tested it on. I am charging 50 dollars per pill. Will you buy it the next time you go out for a night of drinking? Or recommend it to a friend? \u2026 If you said yes, I wonder if you are thinking very critically? Should we think about the cost-benefit ratio here on the basis of what information you have? If you said no, I bet some of the doubts I bring up popped to your mind as well. If 10 out of 15 people saw lessened symptoms, that\u2019s 2\/3 of people \u2013 so some people saw no benefits. Also, what does \u201cgreatly lessened symptoms of hangover\u201d mean? Which symptoms? How much is greatly? Was the reduction by two or more standard deviations from the mean? Or was it less than one standard deviation improvement? Given the cost of 50 dollars per pill, I have to say I would be skeptical about buying it without seeing some statistics!<\/p>\n<p>On this list is a preview of the basic concepts to which you will be introduced as we go through the rest of this chapter.<\/p>\n<div class=\"textbox textbox--learning-objectives\">\n<header class=\"textbox__header\">\n<p class=\"textbox__title\">Hypothesis Testing Basic Concepts<\/p>\n<\/header>\n<div class=\"textbox__content\">\n<ul>\n<li>Hypothesis<\/li>\n<li>Null Hypothesis<\/li>\n<li>Research Hypothesis (alternative hypothesis)<\/li>\n<li>Statistical significance<\/li>\n<li>Conventional levels of significance<\/li>\n<li>Cutoff sample score (critical value)<\/li>\n<li>Directional vs. non-directional hypotheses<\/li>\n<li>One-tailed and two-tailed tests<\/li>\n<li>Type I and Type II errors<\/li>\n<\/ul>\n<\/div>\n<\/div>\n<p>You can see that there are lots of new concepts to master. In my experience, each concept makes the most sense in context, within its place in the <strong>hypothesis testing<\/strong> workflow. We will start with defining our <strong>null<\/strong> and <strong>research hypotheses<\/strong>, then discuss the <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_271\">levels of statistical significance<\/a><\/strong> and their conventional usage. Next, we will look at how to find the <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_270\">cutoff sample score<\/a><\/strong> that will form the critical value for our decision criterion. We will look at how that differs for <strong>directional<\/strong> vs. <strong>non-directional hypotheses<\/strong>, which will lend themselves to <strong>one-<\/strong> or <strong>two-tailed<\/strong> <strong>tests<\/strong>, respectively.<\/p>\n<p>The <strong>hypothesis testing<\/strong> procedure, or workflow, can be broken down into five discrete steps.<\/p>\n<div class=\"textbox textbox--exercises\">\n<header class=\"textbox__header\">\n<p class=\"textbox__title\">Steps of Hypothesis Testing<\/p>\n<\/header>\n<div class=\"textbox__content\">\n<ol>\n<li>Restate question as a research hypothesis and a null hypothesis about populations.<\/li>\n<li>Determine characteristics of the comparison distribution.<\/li>\n<li>Determine the cutoff sample score on the comparison distribution at which the null hypothesis should be rejected.<\/li>\n<li>Determine your sample\u2019s score on the comparison distribution.<\/li>\n<li>Decide whether to reject the null hypothesis.<\/li>\n<\/ol>\n<\/div>\n<\/div>\n<p>These steps are something we will be using pretty much the rest of the semester, so it is worth memorizing them now. My favourite approach to that is to create a mnemonic device. I recommend the following key words from which to form your mnemonic device: hypothesis, characteristics, cutoff, score, and decide. Not very memorable? Try association those with more memorable words that start with the same letter or sound. How about \u201c<em>Happy Chickens Cure Sad Days<\/em>.\u201d Or you can put the words into a <a href=\"https:\/\/www.mnemonicgenerator.com\/\" target=\"_blank\" rel=\"noopener\">mnemonic device generator<\/a> on the internet and get something truly bizarre. I just tried one and got \u201c<em>Hairless Carsick Chewbacca Slapped Demons<\/em>\u201d. Another good one: \u201c<em>Hamlet Chose Cranky Sushi Drunkenly<\/em>.\u201d Anyway, you play around with it or brainstorm until you hit upon one that works for you. Who knew statistics could be this much fun!<\/p>\n<p>The first step in <strong>hypothesis testing<\/strong> is always to formulate hypotheses. The first rule that will help you do so correctly, is that hypotheses are always about <strong>populations<\/strong>. We study samples in order to make conclusions about populations, so our predictions should be about the populations themselves. First, we define <strong>population<\/strong> 1 and <strong>population<\/strong> 2. <strong>Population<\/strong> 1 is always defined as people like the ones in our research study, the ones we are truly interested in. <strong>Population<\/strong> 2 is the comparison <strong>population<\/strong>, the status quo to which we are looking to compare our research <strong>population<\/strong>. Now, remember, when referring to <strong>populations<\/strong>, we always use Greek letters. So if we formulate our hypotheses in symbols, we need to use Greek letters.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-274 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.12-1024x163.png\" alt=\"\" width=\"524\" height=\"83\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.12-1024x163.png 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.12-300x48.png 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.12-768x122.png 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.12-65x10.png 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.12-225x36.png 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.12-350x56.png 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.12.png 1351w\" sizes=\"auto, (max-width: 524px) 100vw, 524px\" \/><\/p>\n<p>It is a good idea to state our hypotheses both in symbols and in words. We need to make them specific and disprovable. If you follow my tips, you will have it down with just a little practice.<\/p>\n<p>We need to state two hypotheses. First, we state the <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_275\">research hypothesis<\/a><\/strong>, which is sometimes referred to as the alternative hypothesis. The <strong>research hypothesis<\/strong> (often called the alternative hypothesis) is a statement of inequality, or that Something happened! This hypothesis makes the prediction that the <strong>population<\/strong> from which the research sample came is different from the comparison <strong>population<\/strong>. In other words, there is a really high <strong>probability<\/strong> that the sample comes from a different distribution than the comparison one.<\/p>\n<p>The <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_277\">null hypothesis<\/a><\/strong>, on the other hand, is a statement of equality, or that nothing happened. This hypothesis makes the prediction that the <strong>population<\/strong> from which sample came is not different from the comparison <strong>population<\/strong>. We set up the <strong>null hypothesis<\/strong> as a so-called straw man, that we hope to tear down. Just remember, null means nothing \u2013 that nothing is different between the <strong>populations<\/strong>.<\/p>\n<p>&nbsp;<\/p>\n<p>Step two of hypothesis testing is to determine the characteristics of the comparison distribution. This is where our descriptive statistics, the mean and standard deviation, come in. We need to ensure our normal curve model to which we are comparing our research <strong>sample<\/strong> is mapped out according to the particular characteristics of the <strong>population<\/strong> of comparison, which is <strong>population<\/strong> 2.<\/p>\n<p>Next it is time to set our decision rule. Step 3 is to determine the <strong>cutoff sample score<\/strong>, which is derived from two pieces of information. The first is the conventional <strong>significance level<\/strong> that applies. By convention, the <strong>probability<\/strong> level that we are willing to accept as a risk that the score from our research sample might occur by random chance within the comparison distribution is set to one of three levels: 10%, 5%, or 1%. The most common choice of <strong>significance level<\/strong> is 5%. Typically the <strong>significance level<\/strong> will be provided to you in the problem for your statistics courses, but if it is not, just default to a <strong>significance level<\/strong> of .05. Sometimes researchers will choose a more conservative <strong>significance level<\/strong>, like 1%, if they are particularly risk averse. If the researcher chooses a 10% <strong>significance level<\/strong>, they are likely conducting a more exploratory study, perhaps a pilot study, and are not too worried about the <strong>probability<\/strong> that the score might be fairly common under the comparison distribution.<\/p>\n<p>The second piece of information we need to know in order to find our <strong>cutoff sample score<\/strong> is which tail we are looking at. Is this a <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_284\">directional hypothesis<\/a><\/strong>, and thus <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_280\">one-tailed test<\/a><\/strong>? Or a <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_285\">non-directional hypothesis<\/a><\/strong>, and thus a <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_281\">two-tailed test<\/a><\/strong>? This depends on the <strong>research hypothesis<\/strong> from step 1. Look for directional keywords in the problem. If the researcher prediction involves words like \u201cgreater than\u201d or \u201clarger than\u201d, this signals that we should be doing a <strong>one-tailed test<\/strong> and that our <strong>cutoff<\/strong> <strong>sample score <\/strong>should be in the top tail of the distribution. If the researcher prediction involves words like \u201clower than\u201d or \u201csmaller than\u201d, this signals that we should be doing a <strong>one-tailed test<\/strong> and that our <strong>cutoff<\/strong> <strong>sample score <\/strong>should be in the bottom tail of the distribution. If the prediction is neutral in directionality, and uses a word like \u201cdifferent\u201d, that signals a <strong>non-directional hypothesis<\/strong>. In that case, we would need to use a two-tailed test, and our cutoff scores would need to be indicated on both tails of the distribution. To do that, we take our area under the curve, which matches our <strong>significance level<\/strong>, and split it into both tails.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-265\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.8-1-1024x558.jpg\" alt=\"\" width=\"554\" height=\"302\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.8-1-1024x558.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.8-1-300x164.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.8-1-768x419.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.8-1-65x35.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.8-1-225x123.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.8-1-350x191.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.8-1.jpg 1537w\" sizes=\"auto, (max-width: 554px) 100vw, 554px\" \/><\/p>\n<p>For example, if we have a <strong>two-tailed test<\/strong> with a .05 <strong>significance level<\/strong>, then we would split the 5% area under the curve into the two tails, so two and a half percent in each tail.<\/p>\n<h4><a id=\"tails_return\"><\/a><a href=\"#tails\">Concept Practice: deciding on one-tailed vs. two-tailed tests<\/a><\/h4>\n<p>We can find the Z-score that forms the border of the tail area we have identified based on <strong>significance level<\/strong> and directionality by looking it up in a <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/back-matter\/normal-curve-z-area-tables\/\" target=\"_blank\" rel=\"noopener\">table<\/a> or an <a href=\"https:\/\/statpages.info\/pdfs.html\" target=\"_blank\" rel=\"noopener\">online calculator<\/a>. I always recommend mapping this <strong>cutoff score<\/strong> onto a drawing of the comparison distribution as shown above. This should help you visualize the setup of the <strong>hypothesis test<\/strong> clearly and accurately.<\/p>\n<h4><a id=\"hypTesting_return\"><\/a><a href=\"#hypTesting\">Concept Practice: inference through hypothesis testing<\/a><\/h4>\n<p>The next step in the <strong>hypothesis testing<\/strong> procedure is to determine your <strong>sample\u2019s<\/strong> score on the comparison distribution. To do this, we calculate a test statistic from the <strong>sample<\/strong> raw score, mark it on the comparison distribution, and determine whether it falls in the shaded tail or not. In reality, we would always have a sample with more than one score in it. However, for the sake of keeping our test statistic formula a familiar one, we will use a sample size of one. We will use our Z-score formula to translate the <strong>sample\u2019s<\/strong> raw score into a Z-score \u2013 in other words, we will figure out how many standard deviations above or below the comparison distribution&#8217;s mean the <strong>sample<\/strong> score is.<\/p>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 37px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-9ff4a1c7099641ee0b6856dfdc537f34_l3.png\" height=\"37\" width=\"95\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#32;&#92;&#091;&#90;&#61;&#92;&#102;&#114;&#97;&#99;&#123;&#88;&#45;&#77;&#125;&#123;&#83;&#68;&#125;&#92;&#093;&#32;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<p>Finally, it\u2019s time to decide whether to <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_288\">reject the null hypothesis<\/a><\/strong>. This decision is based on whether our sample\u2019s data point was more extreme than the <strong>cutoff score<\/strong>, in other words, \u201cdid it fall in the shaded tail?\u201d If the <strong>sample<\/strong> score is more extreme than the <strong>cutoff score<\/strong>, then we must reject the null hypothesis. Our research hypothesis is supported! (Not proven\u2026 remember, there is still some probability that that score could have occurred randomly within the comparison distribution.) But it is sound to say that it appears quite likely that the population from which our sample came is different from the comparison population. Another way to express this decision is to say that the result was <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_291\">statistically significant<\/a><\/strong>, which is to say that there is less than a 5% chance of this result occurring randomly within the comparison distribution (here I just filled in the blank with the significance level).<\/p>\n<p>What if the research sample score did not fall in the shaded tail? In the case that the sample score is less extreme than the <strong>cutoff score<\/strong>, then our <strong>research hypothesis<\/strong> is not supported. We <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_226_289\">do not reject the null hypothesis<\/a><\/strong>. It appears that the <strong>population<\/strong> from which our <strong>sample<\/strong> came is not different from the comparison <strong>population<\/strong>. Note that we do not typically express this result as \u201caccept the null hypothesis\u201d or \u201cwe have proved the null hypothesis\u201d. From this test, we do not have evidence that the <strong>null hypothesis<\/strong> is correct, rather we simply did not have enough evidence to reject it. Another way to express this decision is to say that the result was not <strong>statistically significant<\/strong>, which is to say that there is more than a 5% chance of this result occurring randomly within the comparison distribution (here I just used the most common <strong>significance level<\/strong>).<\/p>\n<h4><a id=\"interpretingHyp_return\"><\/a><a href=\"#interpretingHyp\">Concept Practice: interpreting conclusions of hypothesis tests<\/a><\/h4>\n<p>So we have described the <strong>hypothesis testing<\/strong> process from beginning to end. The whole process of null <strong>hypothesis testing<\/strong> can feel like pretty tortured logic at first. So let us zoom out, and look at the whole process another way. Essentially what we are seeking to do in such a <strong>hypothesis test<\/strong> is to compare two <strong>populations<\/strong>. We want to find out if the <strong>populations<\/strong> are distinct enough to confidently state that there is a difference between <strong>population<\/strong> 1 and <strong>population<\/strong> 2. In our example, we wanted to know if the <strong>population<\/strong> of people using a new medication, <strong>population<\/strong> 1, sleep longer than the <strong>population<\/strong> of people who are not using that new medication, <strong>population<\/strong> 2. We ended up finding that the research evidence to hand suggests <strong>population<\/strong> 1\u2019s distribution is distinct enough from <strong>population<\/strong> 2 that we could <strong>reject the null hypothesis<\/strong> of similarity.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-266 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.9-1-1024x411.jpg\" alt=\"\" width=\"546\" height=\"219\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.9-1-1024x411.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.9-1-300x121.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.9-1-768x309.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.9-1-1536x617.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.9-1-65x26.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.9-1-225x90.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.9-1-350x141.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.9-1.jpg 1974w\" sizes=\"auto, (max-width: 546px) 100vw, 546px\" \/><\/p>\n<p>In other words, we were able to conclude that the difference between the centres of the two distributions was <strong>statistically significant<\/strong>.<\/p>\n<p>If, on the other hand, the distributions were a bit less distinct, we would not have been able to make that claim of a significant difference.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-267 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.10-1-1024x605.jpg\" alt=\"\" width=\"377\" height=\"223\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.10-1-1024x605.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.10-1-300x177.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.10-1-768x454.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.10-1-1536x908.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.10-1-65x38.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.10-1-225x133.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.10-1-350x207.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-4.10-1.jpg 1770w\" sizes=\"auto, (max-width: 377px) 100vw, 377px\" \/><\/p>\n<p>We would not have <strong>rejected the null hypothesis<\/strong> if evidence indicated the <strong>populations<\/strong> were too similar.<\/p>\n<p>Just how different do the two distributions need to be? That criterion is set by the <strong>cutoff score<\/strong>, which depends on the <strong>significance level<\/strong>, and whether it is a <strong>one-tailed<\/strong> or <strong>two-tailed<\/strong> <strong>hypothesis test<\/strong>.<\/p>\n<h4><a id=\"Hyptest_return\"><\/a><a href=\"#Hyptest\">Concept Practice: Putting hypothesis test elements together<\/a><\/h4>\n<p>That was a lot of new concepts to take on! As a reward, assuming you enjoy memes, there are a plethora of <a href=\"https:\/\/www.google.com\/search?q=statistics+memes&amp;rlz=1C1CHBF_enCA784CA784&amp;oq=statistics+memes&amp;aqs=chrome..69i57j0i433i512j0i512l3j69i60l3.2945j1j7&amp;sourceid=chrome&amp;ie=UTF-8\" target=\"_blank\" rel=\"noopener\">statistics memes<\/a>, some of which you may find funny now that you have made it into inferential statistics territory. Welcome to the exclusive club of people who have this rather peculiar sense of humour. Enjoy!<\/p>\n<h1>Chapter Summary<\/h1>\n<p>In this chapter we examined <strong>probability<\/strong> and how it can be used to make inferences about data in the framework of <strong>hypothesis testing<\/strong>. We now have a sense of how two <strong>populations<\/strong> can be compared and the difference between their means evaluated for <strong>statistical significance<\/strong>.<\/p>\n<p>Key Terms:<\/p>\n<table class=\"no-lines\" style=\"border-collapse: collapse;width: 100%;height: 76px\">\n<tbody>\n<tr style=\"height: 15px\">\n<td style=\"width: 27.141%;height: 15px\"><strong>probability<\/strong><\/td>\n<td style=\"width: 29.5125%;height: 15px\"><strong>research hypothesis\u00a0<\/strong><\/td>\n<td style=\"width: 43.8734%;height: 15px\"><strong>one-tailed test<\/strong><\/td>\n<\/tr>\n<tr style=\"height: 15px\">\n<td style=\"width: 27.141%;height: 15px\"><strong>central limit theorem<\/strong><\/td>\n<td style=\"width: 29.5125%;height: 15px\"><strong>null hypothesis <\/strong><\/td>\n<td style=\"width: 43.8734%;height: 15px\"><strong>non-directional hypothesis<\/strong><\/td>\n<\/tr>\n<tr style=\"height: 14px\">\n<td style=\"width: 27.141%;height: 16px\"><strong>population<\/strong><\/td>\n<td style=\"width: 29.5125%;height: 16px\"><strong>cutoff sample score<\/strong><\/td>\n<td style=\"width: 43.8734%;height: 16px\"><strong>statistical significance<\/strong><\/td>\n<\/tr>\n<tr style=\"height: 15px\">\n<td style=\"width: 27.141%;height: 15px\"><strong>sample<\/strong><\/td>\n<td style=\"width: 29.5125%;height: 15px\"><strong>significance level<\/strong><\/td>\n<td style=\"width: 43.8734%;height: 15px\"><strong>reject the null hypothesis<\/strong><\/td>\n<\/tr>\n<tr style=\"height: 15px\">\n<td style=\"width: 27.141%;height: 15px\"><strong>hypothesis testing<\/strong><\/td>\n<td style=\"width: 29.5125%;height: 15px\"><strong>directional hypothesis<\/strong><\/td>\n<td style=\"width: 43.8734%;height: 15px\"><strong>do not<\/strong> <strong>reject the null hypothesis<\/strong><\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<h1>Concept Practice<\/h1>\n<p><a id=\"defineProb\"><\/a><\/p>\n<div id=\"h5p-17\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-17\" class=\"h5p-iframe\" data-content-id=\"17\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4a.02. Probability definition\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#defineProb_return\">text<\/a><\/h6>\n<p><a id=\"calcProb1\"><\/a><\/p>\n<div id=\"h5p-13\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-13\" class=\"h5p-iframe\" data-content-id=\"13\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4a.03. Probability calculation\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#calcProb1_return\">text<\/a><\/h6>\n<p><a id=\"calcProb2\"><\/a><\/p>\n<div id=\"h5p-16\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-16\" class=\"h5p-iframe\" data-content-id=\"16\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4a.06. Probability calculation\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#calcProb2_return\">text<\/a><\/h6>\n<p><a id=\"luckystreak\"><\/a><\/p>\n<div id=\"h5p-15\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-15\" class=\"h5p-iframe\" data-content-id=\"15\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4a.05. Probability calculation\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#lucky_return\">text<\/a><\/h6>\n<p><a id=\"areaProb\"><\/a><\/p>\n<div id=\"h5p-18\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-18\" class=\"h5p-iframe\" data-content-id=\"18\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4a.07 Probability and Z-score (Males&#039; height)\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#areaProb_return\">text<\/a><\/h6>\n<p><a id=\"findPercentile\"><\/a><\/p>\n<div id=\"h5p-19\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-19\" class=\"h5p-iframe\" data-content-id=\"19\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4a.08 Probability and Percentile (Males&#039; height)\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#findPercentile_return\">text<\/a><\/h6>\n<p><a id=\"interpretPercentile\"><\/a><\/p>\n<div id=\"h5p-21\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-21\" class=\"h5p-iframe\" data-content-id=\"21\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4a.10. Probability and Percentile (Blood pressure)\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#interpretPercentile_return\">text<\/a><\/h6>\n<p><a id=\"applyProb\"><\/a><\/p>\n<div id=\"h5p-20\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-20\" class=\"h5p-iframe\" data-content-id=\"20\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4a.09. Probability and Z-score (Blood pressure)\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#applyProb_return\">text<\/a><\/h6>\n<p><a id=\"statsProb\"><\/a><\/p>\n<div id=\"h5p-11\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-11\" class=\"h5p-iframe\" data-content-id=\"11\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4a.01. Probability and Inferential Statistics definition\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#statsProb_return\">text<\/a><\/h6>\n<p><a id=\"CLT\"><\/a><\/p>\n<div id=\"h5p-36\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-36\" class=\"h5p-iframe\" data-content-id=\"36\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5a.07. Central limit theorem\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#CLT_return\">text<\/a><\/h6>\n<p><a id=\"tails\"><\/a><\/p>\n<div id=\"h5p-26\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-26\" class=\"h5p-iframe\" data-content-id=\"26\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4b.05. Non-directional hypothesis true\/false\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#tails_return\">text<\/a><\/h6>\n<p><a id=\"hypTesting\"><\/a><\/p>\n<div id=\"h5p-22\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-22\" class=\"h5p-iframe\" data-content-id=\"22\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4b.01. Hypothesis testing definition\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#hypTesting_return\">text<\/a><\/h6>\n<p><a id=\"interpretingHyp\"><\/a><\/p>\n<div id=\"h5p-23\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-23\" class=\"h5p-iframe\" data-content-id=\"23\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4b.02. Hypothesis testing basic true\/false\"><\/iframe><\/div>\n<\/div>\n<div id=\"h5p-24\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-24\" class=\"h5p-iframe\" data-content-id=\"24\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4b.03. Hypothesis testing basic true\/false\"><\/iframe><\/div>\n<\/div>\n<div id=\"h5p-25\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-25\" class=\"h5p-iframe\" data-content-id=\"25\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4b.04 Hypothesis testing basic true\/false\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#interpretingHyp_return\">text<\/a><\/h6>\n<p><a id=\"Hyptest\"><\/a><\/p>\n<div id=\"h5p-29\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-29\" class=\"h5p-iframe\" data-content-id=\"29\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 4b.05 Hypothesis testing of group leadership\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#Hyptest_return\">text<\/a><\/h6>\n<p>Return to <a href=\"#4a\">4a. Probability and Inferential Statistics<\/a><\/p>\n<h6>Try interactive <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/chapter\/worksheet-4a\/\" target=\"_blank\" rel=\"noopener\">Worksheet 4a<\/a> or download <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/01\/4a.-Class-worksheet.pdf\" target=\"_blank\" rel=\"noopener\">Worksheet 4a<\/a><\/h6>\n<p>Return to <a href=\"#4b\">4b. Hypothesis Testing<\/a><\/p>\n<h6>Try interactive <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/chapter\/worksheet-4b\/\" target=\"_blank\" rel=\"noopener\">Worksheet 4b<\/a>\u00a0or download <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2023\/12\/4b.-Class-worksheet.pdf\">Worksheet 4b<\/a><\/h6>\n<p><a id=\"video4a\"><\/a>video 4a1<\/p>\n<div style=\"width: 1440px;\" class=\"wp-video\"><video class=\"wp-video-shortcode\" id=\"video-226-1\" width=\"1440\" height=\"1080\" preload=\"metadata\" controls=\"controls\"><source type=\"video\/mp4\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/08\/Chapter-4a1.mp4?_=1\" \/><a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/08\/Chapter-4a1.mp4\">https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/08\/Chapter-4a1.mp4<\/a><\/video><\/div>\n<p>video 4a2<\/p>\n<div style=\"width: 1440px;\" class=\"wp-video\"><video class=\"wp-video-shortcode\" id=\"video-226-2\" width=\"1440\" height=\"1080\" preload=\"metadata\" controls=\"controls\"><source type=\"video\/mp4\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Chapter-4a2.mp4?_=2\" \/><a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Chapter-4a2.mp4\">https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Chapter-4a2.mp4<\/a><\/video><\/div>\n<p><a id=\"video4b\"><\/a>video 4b1<\/p>\n<div style=\"width: 1440px;\" class=\"wp-video\"><video class=\"wp-video-shortcode\" id=\"video-226-3\" width=\"1440\" height=\"1080\" preload=\"metadata\" controls=\"controls\"><source type=\"video\/mp4\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Chapter-4b1.mp4?_=3\" \/><a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Chapter-4b1.mp4\">https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Chapter-4b1.mp4<\/a><\/video><\/div>\n<p>video 4b2<\/p>\n<div style=\"width: 1440px;\" class=\"wp-video\"><video class=\"wp-video-shortcode\" id=\"video-226-4\" width=\"1440\" height=\"1080\" preload=\"metadata\" controls=\"controls\"><source type=\"video\/mp4\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/08\/Chapter-4b2.mp4?_=4\" \/><a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/08\/Chapter-4b2.mp4\">https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/08\/Chapter-4b2.mp4<\/a><\/video><\/div>\n<div class=\"glossary\"><span class=\"screen-reader-text\" id=\"definition\">definition<\/span><template id=\"term_226_233\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_233\"><div tabindex=\"-1\"><p>in a situation where several different outcomes are possible, the probability of any specific outcome is a fraction or proportion of all possible outcomes<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_235\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_235\"><div tabindex=\"-1\"><p>mathematical theorem that proposes the following: as long as we take a decent-sized sample, if we took many samples (10,000) of large enough size (30+) and took the mean each time, the distribution of those means will approach a normal distribution, even if the scores from each sample are not normally distributed<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_248\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_248\"><div tabindex=\"-1\"><p>all possible individuals or scores about which we would ideally draw conclusions<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_259\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_259\"><div tabindex=\"-1\"><p>a formal decision making procedure often used in inferential statistics<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_249\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_249\"><div tabindex=\"-1\"><p>the individuals or scores about which we are actually drawing conclusions <\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_271\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_271\"><div tabindex=\"-1\"><p>the probability level that we are willing to accept as a risk that the score from our research sample might occur by random chance within the comparison distribution. By convention, it is set to one of three levels: 10%, 5%, or 1%. <\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_270\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_270\"><div tabindex=\"-1\"><p>critical value that serves as a decision criterion in hypothesis testing<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_275\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_275\"><div tabindex=\"-1\"><p>prediction that the population from which the research sample came is different from the comparison population<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_277\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_277\"><div tabindex=\"-1\"><p>the prediction that the population from which sample came is not different from the comparison population<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_284\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_284\"><div tabindex=\"-1\"><p>a research prediction that the research population mean will be \u201cgreater than\u201d or \"less than\" the comparison population mean<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_280\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_280\"><div tabindex=\"-1\"><p>a hypothesis test in which there is only one cutoff sample score on either the lower or the upper end of the comparison distribution<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_285\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_285\"><div tabindex=\"-1\"><p>a research prediction that the research population mean will be \u201cdifferent from\" the comparison population mean, but allows for the possibility that the research population mean may be either greater than or less than the comparison population mean<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_281\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_281\"><div tabindex=\"-1\"><p>a hypothesis test in which there are two cutoff sample scores, one on either end of the comparison distribution<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_288\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_288\"><div tabindex=\"-1\"><p>a decision in hypothesis testing that concludes statistical significance because the sample score is more extreme than the cutoff score<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_291\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_291\"><div tabindex=\"-1\"><p>the conclusion from a hypothesis test that probability of the observed result occurring randomly within the comparison distribution is less than the significance level<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_226_289\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_226_289\"><div tabindex=\"-1\"><p>a decision in hypothesis testing that is inconclusive because the sample score is less extreme than the cutoff score<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><\/div>","protected":false},"author":1394,"menu_order":4,"template":"","meta":{"pb_show_title":"on","pb_short_title":"","pb_subtitle":"","pb_authors":[],"pb_section_license":""},"chapter-type":[48],"contributor":[],"license":[],"class_list":["post-226","chapter","type-chapter","status-publish","hentry","chapter-type-numberless"],"part":3,"_links":{"self":[{"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/chapters\/226","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/chapters"}],"about":[{"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/wp\/v2\/types\/chapter"}],"author":[{"embeddable":true,"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/wp\/v2\/users\/1394"}],"version-history":[{"count":26,"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/chapters\/226\/revisions"}],"predecessor-version":[{"id":1228,"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/chapters\/226\/revisions\/1228"}],"part":[{"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/parts\/3"}],"metadata":[{"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/chapters\/226\/metadata\/"}],"wp:attachment":[{"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/wp\/v2\/media?parent=226"}],"wp:term":[{"taxonomy":"chapter-type","embeddable":true,"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/chapter-type?post=226"},{"taxonomy":"contributor","embeddable":true,"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/wp\/v2\/contributor?post=226"},{"taxonomy":"license","embeddable":true,"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/wp\/v2\/license?post=226"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}