{"id":318,"date":"2021-08-03T17:41:15","date_gmt":"2021-08-03T21:41:15","guid":{"rendered":"https:\/\/pressbooks.bccampus.ca\/statspsych\/?post_type=chapter&#038;p=318"},"modified":"2023-12-30T19:44:03","modified_gmt":"2023-12-31T00:44:03","slug":"chapter-5","status":"publish","type":"chapter","link":"https:\/\/pressbooks.bccampus.ca\/statspsych\/chapter\/chapter-5\/","title":{"raw":"5. Single Sample Z-test and t-test","rendered":"5. Single Sample Z-test and t-test"},"content":{"raw":"<h1><a id=\"5a\"><\/a>5a. Single Sample Z-test<\/h1>\r\n<h6 style=\"text-align: right\"><a href=\"#video5a\">video lesson<\/a><\/h6>\r\nNow that you have mastered the basic process of hypothesis testing, you are ready for this: your first real statistical test. First, we will examine the types of error that can arise in the context of hypothesis testing.\r\n\r\nWhich type of error is more serious for a professional?\u00a0 1. I decide that a new mode of transportation, the driverless car, is much safer than current modes of transportation, and recommend to the state that they force conversion to that new mode. But then it turns out after a year that the rates of accidents actually increased.\r\n\r\n[caption id=\"\" align=\"aligncenter\" width=\"584\"]<img class=\"\" src=\"https:\/\/live.staticflickr.com\/5592\/15104006386_1bf6bfe96a_b.jpg\" alt=\"Google Self-Driving Car\" width=\"584\" height=\"377\" \/> \"Google Self-Driving Car\" by smoothgroover22 is licensed under CC BY-SA 2.0[\/caption]\r\n\r\nOr... 2. I decide that a new mode of transportation is safer than current modes, but I do not have enough confidence in that margin of safety and therefore do not recommend a sweeping change. After a year it becomes clear that driverless cars are indeed much safer, and a switch would have been beneficial.\r\n\r\nIf you selected the first error as the more serious one, then you chose a <strong>[pb_glossary id=\"320\"]Type I error[\/pb_glossary]<\/strong>. This is the type of error that is considered more serious in the realm of statistical decision making, as well.\r\n\r\nIn the process of hypothesis testing, the decision you make is all about probabilities. It is an educated guess. However, there is always room for error. In fact, there are two types of errors you can make in hypothesis testing. Their imaginative labels are <strong>Type I<\/strong> and <strong>[pb_glossary id=\"321\"]Type II error[\/pb_glossary]<\/strong>. As we saw, <strong>Type I error<\/strong> is considered more serious, so the Significance level is typically set with an eye toward the probability of <strong>Type I error<\/strong> that is deemed acceptable in that study. \u03b1 = .05 indicates that the researcher accepts a 5% chance of a <strong>Type I error<\/strong>. In the decision matrix shown here, you can see how the hypothesis test can play out in four different ways.\r\n\r\n<img class=\"wp-image-319 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-1024x520.png\" alt=\"Decision matrix\" width=\"563\" height=\"286\" \/>\r\n\r\nNote that the real situation, whether the null hypothesis is true or the research hypothesis is true, is unknown. We can never know which is true \u2013 that is why they are hypotheses, after all!\r\n\r\nIf, hypothetically, the null were true, and if we made the decision to reject the null hypothesis, then we would land in this corner of the decision matrix: <strong>Type I error<\/strong>. This is what happens if we go out on a limb, reject the null hypothesis, but we are wrong. As researchers we live in fear of making a <strong>Type I error<\/strong>. No one wants to make a big thing about their research findings and promote change to a new, different, risky, expensive thing, and then be wrong about it. On the other hand, if the null were true and we did not reject it, then we would be correct.\r\n\r\nHypothetically, let us suppose the research hypothesis were true. If we make the decision to reject the null hypothesis, then we are correct. However, if we decide not to reject the null hypothesis, we would be making a <strong>Type II error<\/strong>. This is when we are too conservative, decide there is not enough evidence that the research (alternative) hypothesis could be correct, but we are wrong. <strong>Type II error<\/strong> is also not great, so we do try to minimize that probability by doing something called a power analysis. However, researchers generally would rather make a <strong>Type II<\/strong> than a <strong>Type I<\/strong> error. Keeping the status quo may not do much for to make your scientific career exciting, but objectively speaking it is not such a scary proposition.\r\n\r\nNow, a key thing to remember when you refer back to this decision matrix to determine for which error you are at risk, you never know which column you are in, because the real situation is always unknown. What determines the possible type of error is the thing you have control over: the decision. So you can switch the row you will place yourself in, depending on the observed data, and thus remove yourself from unreasonable risk of <strong>Type I error<\/strong>. (The probability of making a <strong>Type I error<\/strong> is symbolized as <em><strong>[pb_glossary id=\"580\"]\u03b1[\/pb_glossary]<\/strong><\/em>, so significance levels are often referred to as <em><strong>\u03b1 <\/strong><\/em>.)\r\n<h4><a id=\"ErrorTypes_return\"><\/a><a href=\"#ErrorTypes\">Concept Practice: Type I and Type II errors<\/a><\/h4>\r\n&nbsp;\r\n\r\nOkay, we are about to make a connection that is intellectually a bit challenging. So prepare yourself: we will be drawing connections between the central limit theorem, sampling distributions, and the appropriate type of comparison distribution we should use for a hypothesis test where the sample is made up of more than one data point.\r\n\r\nIn the past, we have been using the normal distribution made of up individual scores as our comparison distribution. In the illustration below, if distribution (A) is the distribution of individual scores, (B) is the histogram of a sample of scores drawn from (A). (C) is the <strong>[pb_glossary id=\"328\"]distribution of means[\/pb_glossary]<\/strong> from many samples such as the one portrayed in (B).\r\n\r\n<img class=\"aligncenter wp-image-324 size-large\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.2-1024x390.jpg\" alt=\"Introducing distribution of means\" width=\"1024\" height=\"390\" \/>\r\n<div>The shape of a <strong>distribution of means<\/strong> is narrower than the distribution of individual scores from which it came. The variance is smaller, because it is divided by N, the sample size:<\/div>\r\n<div>[latex] \\[\\sigma^{2}_{M} = \\frac{\\sigma ^{2}}{N}\\] [\/latex]<\/div>\r\n<div><\/div>\r\n<div>The standard deviation is smaller, because it is divided by the square root of N:<\/div>\r\n<div>[latex] \\[\\sigma_{M} = \\frac{\\sigma }{\\sqrt{N}}\\] [\/latex]<\/div>\r\n<div><\/div>\r\n<div>The means are the same for both types of distributions:<\/div>\r\n<div>\r\n\r\n[latex]\\[\\mu_{M} = \\mu\\] [\/latex]\r\n\r\n<\/div>\r\n&nbsp;\r\n<h4><a id=\"DistMeans_return\"><\/a><a href=\"#DistMeans\">Concept Practice: Distribution of Sample Means<\/a><\/h4>\r\n&nbsp;\r\n\r\nUp until now we worked with individuals as our samples, so we could use our old Z-score formula and compare the sample score to a comparison distribution of individuals. We did this, so that we could avoid adding new symbols to our formulas, and concentrate on the hypothesis testing logic. Now, we want the more realistic situation, so we will use the <strong>[pb_glossary id=\"336\"]Z-test[\/pb_glossary]<\/strong> to compare a sample mean to a comparison <strong>distribution of sample means<\/strong>. To rephrase, when we had an individual score, we used the distribution of individuals. When we have a sample mean, we need to use the <strong>distribution of sample means<\/strong>. The new <strong>Z-test<\/strong> statistic formula is shown here:\r\n\r\n[latex] \\[Z=\\frac{M-\\mu _{M}}{\\sigma _{M}}\\] [\/latex]\r\n\r\nIn this formula, M is the sample mean, \u03bc<sub>M<\/sub> is the comparison distribution mean, and \u03c3<sub>M<\/sub> is the comparison distribution standard deviation.\r\n\r\nWe will be using the terminology \u201c<strong>distribution of means<\/strong>,\u201d because it\u2019s a name that reminds us of how it differs from the distribution of individuals. A term in statistics that means the same thing is \u201csampling distribution.\u201d Why do we need this <strong>distribution of means<\/strong>? In reality, we do not compare an individual sample score against a population distribution of individuals. For a real statistical test, we collect a sample and calculate a mean, and compare that sample mean against a comparison\u00a0 distribution of means. We will ground that in an example: If I want to know if my introductory psychology class is collectively performing at the expected level on a test, I do not take their average performance and compare that mean against a set of individual scores, I want to compare it to a set of other class averages. How does this impact our hypothesis test? It matters for Step 2, determining the characteristics of the comparison distribution. Well, it does not change anything about the comparison mean, because the mean of a <strong>distribution of means<\/strong> is equal to the mean of the distribution of individuals. But the standard deviation now needs to be converted into the standard deviation of the <strong>distribution of means<\/strong>, which is often called the <strong>[pb_glossary id=\"334\"]standard error of the mean[\/pb_glossary]<\/strong>. As these equations above show, the variance of a <strong>distribution of means<\/strong> is equal to the variance of the distribution of individuals divided by the number of individuals in the sample. And thus, if we take the square root of the variance, the standard deviation of a <strong>distribution of means<\/strong> is equal to the standard deviation of the distribution of individuals divided by the square root of the number of individuals in the sample.\r\n\r\nWe will now examine the way the steps of the hypothesis test procedure play out when using the single-sample <strong>Z-test<\/strong>. First, we need to formulate the research and null hypotheses. Remember, in such a hypothesis test, we are trying to compare two populations. Population 1, the research population, is the one from which we have drawn a research sample.\r\n\r\n<img class=\"aligncenter wp-image-356\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.3-1.png\" alt=\"\" width=\"499\" height=\"254\" \/>\r\n\r\nWe do not have access to the actual research population mean, hence the need to calculate a sample mean. That sample mean is our best proxy for the population 1 mean, which we need to compare to the mean of population 2. The nature of the comparison, in particular its directionality, is determined by the researcher\u2019s prediction.\r\n\r\nIn step 2, we need to determine the characteristics of the comparison distribution. This means that we need the mean and standard deviation of the comparison distribution, which represents Population 2. This is where we need to introduce the new <strong>distribution of means<\/strong>. The mean is the same as provided, but the standard deviation needs to be converted.\r\n\r\nIf the variance is provided, you can start with this formula to convert to the distribution of means, then take the square root to get the standard deviation:\r\n<div>[latex] \\[\\sigma^{2}_{M} = \\frac{\\sigma ^{2}}{N}\\] [\/latex]<\/div>\r\nIf the standard deviation is provided, it is easiest to use this formula:\r\n\r\n[latex] \\[\\sigma_{M} = \\frac{\\sigma }{\\sqrt{N}}\\] [\/latex]\r\n\r\nStep 3 is to determine the cutoff sample score, which is derived from two pieces of information. First, we need to know the conventional significance level that applies. Second, we need to know which tail we are looking at. The bottom, the top, or both? This depends on the research hypothesis, so we should always look for directional keywords in problem. If the research hypothesis is directional, we should use a one-tailed test. If the research hypothesis is non-directional, we should use a two-tailed test.\r\n\r\nOnce we have those two pieces of information, we can find the Z-score that forms the border of the shaded tail by looking it up in a <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/back-matter\/normal-curve-z-area-tables\/\" target=\"_blank\" rel=\"noopener\">table<\/a>. We can then map the cutoff score onto a drawing of the comparison distribution.\r\n\r\n[caption id=\"attachment_348\" align=\"aligncenter\" width=\"458\"]<img class=\"wp-image-348\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.9-1024x600.jpg\" alt=\"\" width=\"458\" height=\"268\" \/> Example of comparison distribution sketch showing Z-score cutoff and shaded tail representing decision rule[\/caption]\r\n\r\nIn step 4, we need to calculate the Z-test score from the sample mean:\r\n\r\n[latex] \\[Z=\\frac{M-\\mu _{M}}{\\sigma _{M}}\\] [\/latex]\r\n\r\nIn this formula, M represent the research sample mean, <em>\u03bc<sub>M<\/sub><\/em> represents the comparison population mean, and <em>\u03c3<sub>M<\/sub><\/em> represents the standard deviation of the comparison population. Remember, we are using the <strong>distribution of means<\/strong> as our comparison distribution. Once we have calculated the Z-test score, we will mark where it falls on the comparison distribution, and determine whether it falls in the shaded tail or not.\r\n\r\n[caption id=\"attachment_349\" align=\"aligncenter\" width=\"486\"]<img class=\"wp-image-349\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.10-1024x577.jpg\" alt=\"\" width=\"486\" height=\"274\" \/> Example of mapping sample mean (M) onto comparison distribution based on Z-test calculation[\/caption]\r\n\r\nFinally, it is time to decide whether or not to reject the null hypothesis. If the sample mean falls within the shaded tail, we reject the null hypothesis. If it does not, we do not reject the null hypothesis. In other words if the sample mean is extreme enough relative to the comparison distribution of means, then we reject the null hypothesis.\r\n\r\nOnce we make our decision, we will need to take a close look at what kind of error we might have made as a result of our decision. And remember, from our decision matrix, all we have control over whether we're on the top row or the bottom row.\r\n\r\n<img class=\"aligncenter wp-image-319\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-1024x520.png\" alt=\"Decision matrix\" width=\"565\" height=\"287\" \/>\r\n\r\nWe do not know the real situation. if we do make the conclusion that we should reject the null hypothesis, then we are at some risk of a <strong>Type I error<\/strong>, but as long as we don't exceed our significance level, we accept that risk. If we do not reject the null hypothesis we could be at a risk for <strong>Type II error<\/strong>.\r\n\r\nTypically, after conducting a hypothesis test, researchers want to obtain the <strong>[pb_glossary id=\"352\"]p-value[\/pb_glossary]<\/strong>, or the probability of the observed sample score or more extreme occurring just by random chance under the comparison distribution. The p-value associated with the sample mean has to be less than the significance level in order to reject the null hypothesis. A common way to find the <strong>p-value<\/strong> is to use an <a href=\"https:\/\/www.socscistatistics.com\/pvalues\/normaldistribution.aspx\" target=\"_blank\" rel=\"noopener\">online calculator<\/a>.\r\n\r\nOne last exercise I would like you to try is to pretend you are publishing these results in a scientific journal. So you might write this sentence in your results section based on your findings: <span class=\"pullquote-left\">\"We found that patients with insomnia who received a new drug slept significantly more than patients who did not receive the new drug (p &lt; .05).\"<\/span>\r\n\r\nNotice the term \"significantly\" is used in this example, because the null hypothesis was rejected at the end of the hypothesis test. Also, the <strong>p-value<\/strong> is in brackets at the end of this sentence, before the punctuation, just like you would use a citation to back up a factual claim... only here, it is a claim of statistical significance.\r\n\r\n&nbsp;\r\n<h4><a id=\"Ztest_return\"><\/a><a href=\"#Ztest\">Concept Practice: Conducting a Z test<\/a><\/h4>\r\n<h1><a id=\"5b\"><\/a>5b. Single-sample T-test<\/h1>\r\n<h6 style=\"text-align: right\"><a href=\"#video5b\">video lesson<\/a><\/h6>\r\nNow it is time to introduce your second real statistical test: the single-sample <strong>[pb_glossary id=\"358\"]t-test[\/pb_glossary]<\/strong>. This test is much like the <strong>Z-test<\/strong>, but more commonly used in real life.\r\n\r\nBefore we get started on the <strong>t-test<\/strong>, I am curious\u2026 would you have greater confidence in a statistic that comes from a small sample?\r\n\r\n<img class=\"aligncenter wp-image-359\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.6-1024x407.jpg\" alt=\"\" width=\"556\" height=\"221\" \/>\r\n\r\nOr in one that comes from an entire population?... If you said that you would have greater confidence in a number that comes from an entire population, then the field of statistics would agree with you.\r\n\r\nSometimes, though, we do not have access to all the data from population that we need, and we have to take a best guess based on a sample. In that case, we probably want to be more conservative in our decision making. That\u2019s what a <strong>t-test<\/strong> is for.\r\n\r\nOften, even if we have information regarding the comparison population mean, we do not have access to the population standard deviation (<em>\u03c3<\/em>). In this situation, what we must do is collect a sample reflecting population 1, calculate its mean and standard deviation, and compare that sample mean against a comparison <strong>distribution of means<\/strong> using an estimate (<em>S<\/em><sub>M<\/sub>) of the population (<em>\u03c3<sub>M<\/sub><\/em>). Note the change in symbol (from <em>\u03c3 <\/em>to<em> S)<\/em> to reflect the fact that we are basing our estimate on sample data, rather than using a known population parameter.\r\n\r\n[latex] \\[S^{2} = \\frac{\\sum (X-M)^{2}}{N-1}\\] [\/latex]\r\n\r\nYou may have noticed that this new formula for calculating variance in a sample has on the bottom \u201c<em>N<\/em> minus 1\u201d rather than just \u201c<em>N<\/em>\u201d as our old calculation formula had. Why \u201c<em>N<\/em> minus 1\u201d? This is a correction, based on sample size, that derives from the concept of <strong>[pb_glossary id=\"362\"]degrees of freedom[\/pb_glossary]<\/strong>. We will return to that concept later. For now, let us consider \u201c<em>N<\/em> minus 1\u201d as a correction based on sample size that has the following deliberate and intentional effects on our calculation. If we have a small sample size, in which we should not have a great deal of confidence, subtracting 1 really affects the calculation. For example, if we had an <em>N<\/em> of 2, subtracting off 1 from that sample size will mean that we divide by 1 instead of 2. That has a drastic impact on the variance calculation, in effect doubling our estimated variance. Of course in the context of hypothesis testing, more variance is bad \u2013 in that we are less likely to be able to make conclusions based on the data.\r\n\r\nOn the other hand, with a larger sample size, subtracting 1 will have little impact on our calculation\u2026 dividing by 100 or dividing by 99 will have nearly identical results. Think of it this way \u2013 if we have a small sample size, we are punishing ourselves, like a handicap in a golf game. With a larger sample size, we reward ourselves, boosting our chance of a conclusion statistical decision.\r\n\r\nSubtracting off 1 is like a flat tax \u2013 a flat income tax of 10% for all Canadians would be harder on those with a low income, who would have even less money to meet basic survival needs, whereas the wealthiest Canadians would still have plenty of money for their needs. That is, of course, why democratic governments typically use progressive tax rates, because those who have more can afford to pay a greater proportion of their income and still have enough to cover their basic costs of living. With hypothesis testing, though, the flat tax is desirable, because it is our intent to penalize the smallest sample sizes the most, given the fact that we can have little confidence in the estimates we derive from them.\r\n\r\nThe t-test is designed around the concept of <strong>degrees of freedom<\/strong>, which is defined as the number of scores in a given calculation that are free to vary. This is an abstract definition that is difficult to grasp, so it is helpful to consider a concrete example.\r\n<div class=\"textbox textbox--examples\"><header class=\"textbox__header\">\r\n<p class=\"textbox__title\">Degrees of Freedom Practical Example<\/p>\r\n\r\n<\/header>\r\n<div class=\"textbox__content\">\r\n<blockquote>First, consider the game of baseball.\u00a0 We understand the field-of-play consists of 9 positions.\u00a0 The coach is \u201cfree\u201d to assign any of the 9 players to any of the 9 positions. Once the 8th player is assigned to the 8th position, the 9th player-position is pre-determined, so to speak.\u00a0 In other words, the coach is not \u201cfree\u201d to pick either the last position or the last player.<\/blockquote>\r\n<em>Source:<\/em> https:\/\/www.isixsigma.com\/ask-dr-mikel-harry\/ask-tools-techniques\/can-you-explain-degrees-freedom-and-provide-example\/\r\n\r\n<\/div>\r\n<\/div>\r\nThe same logic holds when doing a calculation involving all of a sample\u2019s scores. Up until the last score inputted into the formula, the scores are free to vary; but the last score is pre-determined. Thus <strong>degrees of freedom<\/strong> in a <strong>t-test<\/strong> is <em>N<\/em>-1, or 1 less than the total number of scores in the dataset.\r\n\r\nIn a <strong>t-test<\/strong> we will be using the <strong>t-distributions<\/strong>, rather than a normal distribution, as our comparison distributions. The <strong>t-distributions<\/strong> are a series of distributions, based on the normal distribution, that adjust their shape according to <strong>degrees of freedom<\/strong> (which in turn is based on sample size).\r\n\r\nAs mentioned previously, when we do not have the population standard deviation we must estimate it from our sample. This involves a greater risk of error. The risk is related to how large our sample is. So, the fewer the scores in our sample, the fewer the degrees of freedom, and the more conservative the distribution we must use \u2013 one that is wider and shorter (more area in the tails).\r\n\r\n<img class=\"aligncenter wp-image-365\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.7-1024x566.jpg\" alt=\"\" width=\"455\" height=\"251\" \/>\r\n\r\nAbove you can see examples of how the shape of the t-distributions changes with reduced <strong>degrees of freedom<\/strong>. With <strong>degrees of freedom<\/strong> of 100, the shape of the distribution is very close to the normal curve. If we have very few <strong>degrees of freedom<\/strong>, like 1, then our shading rejection regions, or goal posts, will move farther out into the tails of the distribution, because there is more area under the tails of the curve. In effect, then, with fewer <strong>degrees of freedom<\/strong>, we will find it more difficult to reject the null hypothesis. This is, of course, by design, in order to penalize the hypothesis test if it is based on a very small sample size.\r\n\r\n[caption id=\"attachment_366\" align=\"aligncenter\" width=\"484\"]<img class=\"wp-image-366\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.8-1024x609.jpg\" alt=\"\" width=\"484\" height=\"288\" \/> <em>Inspiration for this creative illustration came from one of my former students, Andi.\u00a0Blame for artwork quality is entirely on me.<\/em>[\/caption]\r\n\r\nAbove is a less technical visual to see the difference in distribution shape when the sample size is small, versus when the sample size is large. As you can see, when you have a large sample size, the goal posts are very close and easy to reach, but when you have a small sample size, they will really be far away, and this will make it difficult to reject our null hypothesis.\r\n<h4><a id=\"Ttest_return\"><\/a><a href=\"#Ttest\">Concept Practice: Sample size and the t-test<\/a><\/h4>\r\n&nbsp;\r\n\r\nNow we are ready to take a look at how a hypothesis test works based on a single-sample <strong>t-test<\/strong>. As before, the structure of the test is familiar \u2013 we are comparing the means of Population 1 and Population 2.\r\n\r\n<img class=\"aligncenter wp-image-369\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.11.png\" alt=\"\" width=\"532\" height=\"325\" \/>\r\n\r\nAs usual, we do not have access to the mean of population 1, just the mean of a sample from that population. Furthermore, unlike in a <strong>Z-test<\/strong>, we do not have access to one piece of information about population 2: the population standard deviation. Therefore, we will form hypotheses the same way, but when we do steps 2 and 4, we will need to use the sample to generate an estimate of the comparison population standard deviation.\r\n<h4><a id=\"ZorT_return\"><\/a><a href=\"#ZorT\">Concept Practice: When to use a Z-test or a t-test<\/a><\/h4>\r\n&nbsp;\r\n\r\nIn step 2, we identify the known comparison population\u2019s mean. However, we will need to use the sample to generate an estimate of the comparison population standard deviation. We calculate the sample variance, using the formula with the \u201c<em>N<\/em>-1\u201d correction,\r\n\r\n[latex] \\[S^{2} = \\frac{\\sum (X-M)^{2}}{N-1}\\] [\/latex]\r\n\r\nand then convert the variance to the <strong>distribution of means<\/strong> by dividing by <em>N<\/em>,\r\n\r\n[latex] \\[S^{2}_{M} = \\frac{S ^{2}}{N}\\] [\/latex]\r\n\r\nand finally we square root to get from variance to standard deviation.\r\n\r\n[latex] \\[S_{M} = \\sqrt{S^{2}_{M}}\\][\/latex]\r\n\r\n<em>S<sub>M<\/sub><\/em> is what we will use to describe our comparison distribution.\r\n\r\nIn step 3, as before we need to find our cutoff\u00a0 sample score based on the significance level and directionality of the test. Now, however, we also need to use a third piece of information: <strong>degrees of freedom<\/strong>. This is found by subtracting one from the sample size. Then we can look in the <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/back-matter\/t-distribution-tables\/\" target=\"_blank\" rel=\"noopener\">t-tables<\/a>, identifying the row of the table with the matching <strong>degrees of freedom<\/strong>, then looking in the appropriate column based on directionality and significance level. Once we identify the t-score that forms the border of the shaded tail, we can map this onto a drawing of the comparison distribution. For visualization purposes, it is fine to sketch the distribution the same way you did the normal distribution.\r\n<h4><a id=\"DF_return\"><\/a><a href=\"#DF\">Concept Practice: Degrees of Freedom<\/a><\/h4>\r\n&nbsp;\r\n\r\n&nbsp;\r\n\r\nNow for step 4. The <strong>t-test<\/strong> formula is just like the <strong>Z-test<\/strong> one on top \u2013 sample mean minus comparison population mean. Then you divide by <em>S<sub>M<\/sub><\/em>, the estimate of standard deviation that you calculated back in step 2.\r\n\r\n[latex] \\[t = \\frac{M-\\mu _{M}}{S_{M}}\\][\/latex]\r\n\r\nIf your calculated <strong>t-test<\/strong> score fell into the shaded tail beyond your cutoff score, then you may reject the null hypothesis. In other words, if the sample mean was extreme enough on the comparison distribution of means, reject the null.\r\n\r\n<img class=\"aligncenter size-large wp-image-377\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/A594CFBD-6310-4E2B-8121-9C3B91C5E9F2-1024x616.jpeg\" alt=\"\" width=\"1024\" height=\"616\" \/>\r\n\r\nAfter conducting a hypothesis test, it is helpful to also obtain a report the precise <strong>p-value<\/strong> associated with the result. The <strong>p-value<\/strong> offers the precise probability of obtaining the <strong>t-test<\/strong> result (or more extreme) just by random chance under the comparison distribution. It is found by <a href=\"https:\/\/www.socscistatistics.com\/pvalues\/tdistribution.aspx\" target=\"_blank\" rel=\"noopener\">calculating<\/a> the area under the comparison distribution in the tail beyond the observed <strong>t-test<\/strong> score calculated in step 4 (or in both tails, if it was a two-tailed test).\r\n\r\n&nbsp;\r\n<h4><a id=\"tTest_return\"><\/a><a href=\"#tTest\">Concept Practice: Conduct a t-test<\/a><\/h4>\r\n<h4><a id=\"Facts_return\"><\/a><a href=\"#Facts\">Concept Practice: Facts about Z- and t-tests<\/a><\/h4>\r\n&nbsp;\r\n\r\n&nbsp;\r\n<h1>Chapter Summary<\/h1>\r\nIn this chapter we learned about the two types of logical errors we can make in hypothesis testing: <strong>Type I<\/strong> and <strong>Type II errors<\/strong>. We saw that <strong>Type I error<\/strong> is considered more serious in statistics, because it represents going out on limb, making a strong conclusion, when that is in fact the wrong decision. For that reason, <strong>Type I error<\/strong> risk is strictly controlled in inferential statistics by setting a particular significance level (<em><strong>\u03b1<\/strong><\/em>). We also introduced our first real inferential statistical tests: the single-sample <strong>Z-test<\/strong> and the single-sample <strong>t-test<\/strong>. In each of these tests, we are comparing the means of two populations, using a sample to estimate the mean of the research population. A <strong>Z-test<\/strong> is used when we know the standard deviation of the comparison population (<em><strong>\u03c3<\/strong><\/em>); a <strong>t-test<\/strong> is used when we do not have that information and must estimate the standard deviation from the sample (<strong><em>S<\/em><\/strong>).\r\n\r\nKey terms:\r\n<table class=\"no-lines\" style=\"border-collapse: collapse;width: 100%\" border=\"0\">\r\n<tbody>\r\n<tr>\r\n<td style=\"width: 33.3333%\"><strong>Type I error<\/strong><\/td>\r\n<td style=\"width: 33.3333%\"><strong>distribution of means<\/strong><\/td>\r\n<td style=\"width: 33.3333%\"><strong>p-value<\/strong><\/td>\r\n<\/tr>\r\n<tr>\r\n<td style=\"width: 33.3333%\"><strong>Type II error<\/strong><\/td>\r\n<td style=\"width: 33.3333%\"><strong>Z-test<\/strong><\/td>\r\n<td style=\"width: 33.3333%\"><strong>t-test<\/strong><\/td>\r\n<\/tr>\r\n<tr>\r\n<td style=\"width: 33.3333%\"><em><strong>\u03b1<\/strong><\/em><\/td>\r\n<td style=\"width: 33.3333%\"><strong>standard error of the mean<\/strong><\/td>\r\n<td style=\"width: 33.3333%\"><strong>degrees of freedom<\/strong><\/td>\r\n<\/tr>\r\n<\/tbody>\r\n<\/table>\r\n<h1>Concept Practice<\/h1>\r\n<a id=\"ErrorTypes\"><\/a>[h5p id=\"31\"]\r\n\r\n[h5p id=\"32\"]\r\n\r\n[h5p id=\"33\"]\r\n\r\n[h5p id=\"34\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#ErrorTypes_return\">text<\/a><\/h6>\r\n<a id=\"DistMeans\"><\/a>[h5p id=\"37\"]\r\n\r\n[h5p id=\"38\"]\r\n\r\n[h5p id=\"39\"]\r\n\r\n[h5p id=\"40\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#DistMeans_return\">text<\/a><\/h6>\r\n<a id=\"Ztest\"><\/a>[h5p id=\"41\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#Ztest_return\">text<\/a><\/h6>\r\n<a id=\"Ttest\"><\/a>[h5p id=\"43\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#Ttest_return\">text<\/a><\/h6>\r\n<a id=\"ZorT\"><\/a>[h5p id=\"44\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#ZorT_return\">text<\/a><\/h6>\r\n<a id=\"DF\"><\/a>[h5p id=\"42\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#DF_return\">text<\/a><\/h6>\r\n<a id=\"tTest\"><\/a>[h5p id=\"46\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#tTest_return\">text<\/a><\/h6>\r\n<a id=\"Facts\"><\/a>[h5p id=\"45\"]\r\n<h6 style=\"text-align: right\">Return to <a href=\"#Facts_return\">text<\/a><\/h6>\r\nReturn to <a href=\"#5a\">5a. Single-sample Z test<\/a>\r\n<h6>Try interactive <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/?post_type=chapter&amp;p=889&amp;preview=true\">Worksheet 5a<\/a> or download Worksheet 5a<\/h6>\r\nReturn to <a href=\"#5b\">5b. Single-sample t test<\/a>\r\n<h6>Try interactive Worksheet 5b\u00a0or download Worksheet 5b<\/h6>\r\n&nbsp;\r\n\r\n<a id=\"video5a\"><\/a>video 5a1\r\n\r\n[video width=\"1440\" height=\"1080\" mp4=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/08\/Chapter-5a1.mp4\"][\/video]\r\n\r\nvideo 5a2\r\n\r\n&nbsp;\r\n\r\n<a id=\"video5b\"><\/a>video 5b1","rendered":"<h1><a id=\"5a\"><\/a>5a. Single Sample Z-test<\/h1>\n<h6 style=\"text-align: right\"><a href=\"#video5a\">video lesson<\/a><\/h6>\n<p>Now that you have mastered the basic process of hypothesis testing, you are ready for this: your first real statistical test. First, we will examine the types of error that can arise in the context of hypothesis testing.<\/p>\n<p>Which type of error is more serious for a professional?\u00a0 1. I decide that a new mode of transportation, the driverless car, is much safer than current modes of transportation, and recommend to the state that they force conversion to that new mode. But then it turns out after a year that the rates of accidents actually increased.<\/p>\n<figure style=\"width: 584px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"\" src=\"https:\/\/live.staticflickr.com\/5592\/15104006386_1bf6bfe96a_b.jpg\" alt=\"Google Self-Driving Car\" width=\"584\" height=\"377\" \/><figcaption class=\"wp-caption-text\">&#8220;Google Self-Driving Car&#8221; by smoothgroover22 is licensed under CC BY-SA 2.0<\/figcaption><\/figure>\n<p>Or&#8230; 2. I decide that a new mode of transportation is safer than current modes, but I do not have enough confidence in that margin of safety and therefore do not recommend a sweeping change. After a year it becomes clear that driverless cars are indeed much safer, and a switch would have been beneficial.<\/p>\n<p>If you selected the first error as the more serious one, then you chose a <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_318_320\">Type I error<\/a><\/strong>. This is the type of error that is considered more serious in the realm of statistical decision making, as well.<\/p>\n<p>In the process of hypothesis testing, the decision you make is all about probabilities. It is an educated guess. However, there is always room for error. In fact, there are two types of errors you can make in hypothesis testing. Their imaginative labels are <strong>Type I<\/strong> and <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_318_321\">Type II error<\/a><\/strong>. As we saw, <strong>Type I error<\/strong> is considered more serious, so the Significance level is typically set with an eye toward the probability of <strong>Type I error<\/strong> that is deemed acceptable in that study. \u03b1 = .05 indicates that the researcher accepts a 5% chance of a <strong>Type I error<\/strong>. In the decision matrix shown here, you can see how the hypothesis test can play out in four different ways.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-319 aligncenter\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-1024x520.png\" alt=\"Decision matrix\" width=\"563\" height=\"286\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-1024x520.png 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-300x152.png 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-768x390.png 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-65x33.png 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-225x114.png 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-350x178.png 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1.png 1042w\" sizes=\"auto, (max-width: 563px) 100vw, 563px\" \/><\/p>\n<p>Note that the real situation, whether the null hypothesis is true or the research hypothesis is true, is unknown. We can never know which is true \u2013 that is why they are hypotheses, after all!<\/p>\n<p>If, hypothetically, the null were true, and if we made the decision to reject the null hypothesis, then we would land in this corner of the decision matrix: <strong>Type I error<\/strong>. This is what happens if we go out on a limb, reject the null hypothesis, but we are wrong. As researchers we live in fear of making a <strong>Type I error<\/strong>. No one wants to make a big thing about their research findings and promote change to a new, different, risky, expensive thing, and then be wrong about it. On the other hand, if the null were true and we did not reject it, then we would be correct.<\/p>\n<p>Hypothetically, let us suppose the research hypothesis were true. If we make the decision to reject the null hypothesis, then we are correct. However, if we decide not to reject the null hypothesis, we would be making a <strong>Type II error<\/strong>. This is when we are too conservative, decide there is not enough evidence that the research (alternative) hypothesis could be correct, but we are wrong. <strong>Type II error<\/strong> is also not great, so we do try to minimize that probability by doing something called a power analysis. However, researchers generally would rather make a <strong>Type II<\/strong> than a <strong>Type I<\/strong> error. Keeping the status quo may not do much for to make your scientific career exciting, but objectively speaking it is not such a scary proposition.<\/p>\n<p>Now, a key thing to remember when you refer back to this decision matrix to determine for which error you are at risk, you never know which column you are in, because the real situation is always unknown. What determines the possible type of error is the thing you have control over: the decision. So you can switch the row you will place yourself in, depending on the observed data, and thus remove yourself from unreasonable risk of <strong>Type I error<\/strong>. (The probability of making a <strong>Type I error<\/strong> is symbolized as <em><strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_318_580\">\u03b1<\/a><\/strong><\/em>, so significance levels are often referred to as <em><strong>\u03b1 <\/strong><\/em>.)<\/p>\n<h4><a id=\"ErrorTypes_return\"><\/a><a href=\"#ErrorTypes\">Concept Practice: Type I and Type II errors<\/a><\/h4>\n<p>&nbsp;<\/p>\n<p>Okay, we are about to make a connection that is intellectually a bit challenging. So prepare yourself: we will be drawing connections between the central limit theorem, sampling distributions, and the appropriate type of comparison distribution we should use for a hypothesis test where the sample is made up of more than one data point.<\/p>\n<p>In the past, we have been using the normal distribution made of up individual scores as our comparison distribution. In the illustration below, if distribution (A) is the distribution of individual scores, (B) is the histogram of a sample of scores drawn from (A). (C) is the <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_318_328\">distribution of means<\/a><\/strong> from many samples such as the one portrayed in (B).<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-324 size-large\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.2-1024x390.jpg\" alt=\"Introducing distribution of means\" width=\"1024\" height=\"390\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.2-1024x390.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.2-300x114.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.2-768x292.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.2-1536x585.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.2-2048x780.jpg 2048w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.2-65x25.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.2-225x86.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.2-350x133.jpg 350w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/p>\n<div>The shape of a <strong>distribution of means<\/strong> is narrower than the distribution of individual scores from which it came. The variance is smaller, because it is divided by N, the sample size:<\/div>\n<div>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 39px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-dddc6771e8e0567c6019f9e5447841fe_l3.png\" height=\"39\" width=\"70\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#32;&#92;&#091;&#92;&#115;&#105;&#103;&#109;&#97;&#94;&#123;&#50;&#125;&#95;&#123;&#77;&#125;&#32;&#61;&#32;&#92;&#102;&#114;&#97;&#99;&#123;&#92;&#115;&#105;&#103;&#109;&#97;&#32;&#94;&#123;&#50;&#125;&#125;&#123;&#78;&#125;&#92;&#093;&#32;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<\/div>\n<div><\/div>\n<div>The standard deviation is smaller, because it is divided by the square root of N:<\/div>\n<div>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 36px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-79feb6b9641740458aca68473e845118_l3.png\" height=\"36\" width=\"83\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#32;&#92;&#091;&#92;&#115;&#105;&#103;&#109;&#97;&#95;&#123;&#77;&#125;&#32;&#61;&#32;&#92;&#102;&#114;&#97;&#99;&#123;&#92;&#115;&#105;&#103;&#109;&#97;&#32;&#125;&#123;&#92;&#115;&#113;&#114;&#116;&#123;&#78;&#125;&#125;&#92;&#093;&#32;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<\/div>\n<div><\/div>\n<div>The means are the same for both types of distributions:<\/div>\n<div>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 12px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-fc0b6112a05ecfa074076c4065a6129f_l3.png\" height=\"12\" width=\"60\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#92;&#091;&#92;&#109;&#117;&#95;&#123;&#77;&#125;&#32;&#61;&#32;&#92;&#109;&#117;&#92;&#093;&#32;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<\/div>\n<p>&nbsp;<\/p>\n<h4><a id=\"DistMeans_return\"><\/a><a href=\"#DistMeans\">Concept Practice: Distribution of Sample Means<\/a><\/h4>\n<p>&nbsp;<\/p>\n<p>Up until now we worked with individuals as our samples, so we could use our old Z-score formula and compare the sample score to a comparison distribution of individuals. We did this, so that we could avoid adding new symbols to our formulas, and concentrate on the hypothesis testing logic. Now, we want the more realistic situation, so we will use the <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_318_336\">Z-test<\/a><\/strong> to compare a sample mean to a comparison <strong>distribution of sample means<\/strong>. To rephrase, when we had an individual score, we used the distribution of individuals. When we have a sample mean, we need to use the <strong>distribution of sample means<\/strong>. The new <strong>Z-test<\/strong> statistic formula is shown here:<\/p>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 39px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-441e8a64eccfe6c4e748fb4c2a49b5f0_l3.png\" height=\"39\" width=\"105\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#32;&#92;&#091;&#90;&#61;&#92;&#102;&#114;&#97;&#99;&#123;&#77;&#45;&#92;&#109;&#117;&#32;&#95;&#123;&#77;&#125;&#125;&#123;&#92;&#115;&#105;&#103;&#109;&#97;&#32;&#95;&#123;&#77;&#125;&#125;&#92;&#093;&#32;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<p>In this formula, M is the sample mean, \u03bc<sub>M<\/sub> is the comparison distribution mean, and \u03c3<sub>M<\/sub> is the comparison distribution standard deviation.<\/p>\n<p>We will be using the terminology \u201c<strong>distribution of means<\/strong>,\u201d because it\u2019s a name that reminds us of how it differs from the distribution of individuals. A term in statistics that means the same thing is \u201csampling distribution.\u201d Why do we need this <strong>distribution of means<\/strong>? In reality, we do not compare an individual sample score against a population distribution of individuals. For a real statistical test, we collect a sample and calculate a mean, and compare that sample mean against a comparison\u00a0 distribution of means. We will ground that in an example: If I want to know if my introductory psychology class is collectively performing at the expected level on a test, I do not take their average performance and compare that mean against a set of individual scores, I want to compare it to a set of other class averages. How does this impact our hypothesis test? It matters for Step 2, determining the characteristics of the comparison distribution. Well, it does not change anything about the comparison mean, because the mean of a <strong>distribution of means<\/strong> is equal to the mean of the distribution of individuals. But the standard deviation now needs to be converted into the standard deviation of the <strong>distribution of means<\/strong>, which is often called the <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_318_334\">standard error of the mean<\/a><\/strong>. As these equations above show, the variance of a <strong>distribution of means<\/strong> is equal to the variance of the distribution of individuals divided by the number of individuals in the sample. And thus, if we take the square root of the variance, the standard deviation of a <strong>distribution of means<\/strong> is equal to the standard deviation of the distribution of individuals divided by the square root of the number of individuals in the sample.<\/p>\n<p>We will now examine the way the steps of the hypothesis test procedure play out when using the single-sample <strong>Z-test<\/strong>. First, we need to formulate the research and null hypotheses. Remember, in such a hypothesis test, we are trying to compare two populations. Population 1, the research population, is the one from which we have drawn a research sample.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-356\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.3-1.png\" alt=\"\" width=\"499\" height=\"254\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.3-1.png 1021w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.3-1-300x153.png 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.3-1-768x392.png 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.3-1-65x33.png 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.3-1-225x115.png 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.3-1-350x179.png 350w\" sizes=\"auto, (max-width: 499px) 100vw, 499px\" \/><\/p>\n<p>We do not have access to the actual research population mean, hence the need to calculate a sample mean. That sample mean is our best proxy for the population 1 mean, which we need to compare to the mean of population 2. The nature of the comparison, in particular its directionality, is determined by the researcher\u2019s prediction.<\/p>\n<p>In step 2, we need to determine the characteristics of the comparison distribution. This means that we need the mean and standard deviation of the comparison distribution, which represents Population 2. This is where we need to introduce the new <strong>distribution of means<\/strong>. The mean is the same as provided, but the standard deviation needs to be converted.<\/p>\n<p>If the variance is provided, you can start with this formula to convert to the distribution of means, then take the square root to get the standard deviation:<\/p>\n<div>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 39px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-dddc6771e8e0567c6019f9e5447841fe_l3.png\" height=\"39\" width=\"70\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#32;&#92;&#091;&#92;&#115;&#105;&#103;&#109;&#97;&#94;&#123;&#50;&#125;&#95;&#123;&#77;&#125;&#32;&#61;&#32;&#92;&#102;&#114;&#97;&#99;&#123;&#92;&#115;&#105;&#103;&#109;&#97;&#32;&#94;&#123;&#50;&#125;&#125;&#123;&#78;&#125;&#92;&#093;&#32;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<\/div>\n<p>If the standard deviation is provided, it is easiest to use this formula:<\/p>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 36px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-79feb6b9641740458aca68473e845118_l3.png\" height=\"36\" width=\"83\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#32;&#92;&#091;&#92;&#115;&#105;&#103;&#109;&#97;&#95;&#123;&#77;&#125;&#32;&#61;&#32;&#92;&#102;&#114;&#97;&#99;&#123;&#92;&#115;&#105;&#103;&#109;&#97;&#32;&#125;&#123;&#92;&#115;&#113;&#114;&#116;&#123;&#78;&#125;&#125;&#92;&#093;&#32;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<p>Step 3 is to determine the cutoff sample score, which is derived from two pieces of information. First, we need to know the conventional significance level that applies. Second, we need to know which tail we are looking at. The bottom, the top, or both? This depends on the research hypothesis, so we should always look for directional keywords in problem. If the research hypothesis is directional, we should use a one-tailed test. If the research hypothesis is non-directional, we should use a two-tailed test.<\/p>\n<p>Once we have those two pieces of information, we can find the Z-score that forms the border of the shaded tail by looking it up in a <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/back-matter\/normal-curve-z-area-tables\/\" target=\"_blank\" rel=\"noopener\">table<\/a>. We can then map the cutoff score onto a drawing of the comparison distribution.<\/p>\n<figure id=\"attachment_348\" aria-describedby=\"caption-attachment-348\" style=\"width: 458px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-348\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.9-1024x600.jpg\" alt=\"\" width=\"458\" height=\"268\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.9-1024x600.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.9-300x176.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.9-768x450.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.9-1536x900.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.9-65x38.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.9-225x132.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.9-350x205.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.9.jpg 1580w\" sizes=\"auto, (max-width: 458px) 100vw, 458px\" \/><figcaption id=\"caption-attachment-348\" class=\"wp-caption-text\">Example of comparison distribution sketch showing Z-score cutoff and shaded tail representing decision rule<\/figcaption><\/figure>\n<p>In step 4, we need to calculate the Z-test score from the sample mean:<\/p>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 39px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-441e8a64eccfe6c4e748fb4c2a49b5f0_l3.png\" height=\"39\" width=\"105\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#32;&#92;&#091;&#90;&#61;&#92;&#102;&#114;&#97;&#99;&#123;&#77;&#45;&#92;&#109;&#117;&#32;&#95;&#123;&#77;&#125;&#125;&#123;&#92;&#115;&#105;&#103;&#109;&#97;&#32;&#95;&#123;&#77;&#125;&#125;&#92;&#093;&#32;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<p>In this formula, M represent the research sample mean, <em>\u03bc<sub>M<\/sub><\/em> represents the comparison population mean, and <em>\u03c3<sub>M<\/sub><\/em> represents the standard deviation of the comparison population. Remember, we are using the <strong>distribution of means<\/strong> as our comparison distribution. Once we have calculated the Z-test score, we will mark where it falls on the comparison distribution, and determine whether it falls in the shaded tail or not.<\/p>\n<figure id=\"attachment_349\" aria-describedby=\"caption-attachment-349\" style=\"width: 486px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-349\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.10-1024x577.jpg\" alt=\"\" width=\"486\" height=\"274\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.10-1024x577.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.10-300x169.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.10-768x433.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.10-1536x865.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.10-65x37.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.10-225x127.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.10-350x197.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.10.jpg 1592w\" sizes=\"auto, (max-width: 486px) 100vw, 486px\" \/><figcaption id=\"caption-attachment-349\" class=\"wp-caption-text\">Example of mapping sample mean (M) onto comparison distribution based on Z-test calculation<\/figcaption><\/figure>\n<p>Finally, it is time to decide whether or not to reject the null hypothesis. If the sample mean falls within the shaded tail, we reject the null hypothesis. If it does not, we do not reject the null hypothesis. In other words if the sample mean is extreme enough relative to the comparison distribution of means, then we reject the null hypothesis.<\/p>\n<p>Once we make our decision, we will need to take a close look at what kind of error we might have made as a result of our decision. And remember, from our decision matrix, all we have control over whether we&#8217;re on the top row or the bottom row.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-319\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-1024x520.png\" alt=\"Decision matrix\" width=\"565\" height=\"287\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-1024x520.png 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-300x152.png 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-768x390.png 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-65x33.png 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-225x114.png 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1-350x178.png 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.1.png 1042w\" sizes=\"auto, (max-width: 565px) 100vw, 565px\" \/><\/p>\n<p>We do not know the real situation. if we do make the conclusion that we should reject the null hypothesis, then we are at some risk of a <strong>Type I error<\/strong>, but as long as we don&#8217;t exceed our significance level, we accept that risk. If we do not reject the null hypothesis we could be at a risk for <strong>Type II error<\/strong>.<\/p>\n<p>Typically, after conducting a hypothesis test, researchers want to obtain the <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_318_352\">p-value<\/a><\/strong>, or the probability of the observed sample score or more extreme occurring just by random chance under the comparison distribution. The p-value associated with the sample mean has to be less than the significance level in order to reject the null hypothesis. A common way to find the <strong>p-value<\/strong> is to use an <a href=\"https:\/\/www.socscistatistics.com\/pvalues\/normaldistribution.aspx\" target=\"_blank\" rel=\"noopener\">online calculator<\/a>.<\/p>\n<p>One last exercise I would like you to try is to pretend you are publishing these results in a scientific journal. So you might write this sentence in your results section based on your findings: <span class=\"pullquote-left\">&#8220;We found that patients with insomnia who received a new drug slept significantly more than patients who did not receive the new drug (p &lt; .05).&#8221;<\/span><\/p>\n<p>Notice the term &#8220;significantly&#8221; is used in this example, because the null hypothesis was rejected at the end of the hypothesis test. Also, the <strong>p-value<\/strong> is in brackets at the end of this sentence, before the punctuation, just like you would use a citation to back up a factual claim&#8230; only here, it is a claim of statistical significance.<\/p>\n<p>&nbsp;<\/p>\n<h4><a id=\"Ztest_return\"><\/a><a href=\"#Ztest\">Concept Practice: Conducting a Z test<\/a><\/h4>\n<h1><a id=\"5b\"><\/a>5b. Single-sample T-test<\/h1>\n<h6 style=\"text-align: right\"><a href=\"#video5b\">video lesson<\/a><\/h6>\n<p>Now it is time to introduce your second real statistical test: the single-sample <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_318_358\">t-test<\/a><\/strong>. This test is much like the <strong>Z-test<\/strong>, but more commonly used in real life.<\/p>\n<p>Before we get started on the <strong>t-test<\/strong>, I am curious\u2026 would you have greater confidence in a statistic that comes from a small sample?<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-359\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.6-1024x407.jpg\" alt=\"\" width=\"556\" height=\"221\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.6-1024x407.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.6-300x119.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.6-768x305.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.6-1536x610.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.6-65x26.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.6-225x89.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.6-350x139.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.6.jpg 1948w\" sizes=\"auto, (max-width: 556px) 100vw, 556px\" \/><\/p>\n<p>Or in one that comes from an entire population?&#8230; If you said that you would have greater confidence in a number that comes from an entire population, then the field of statistics would agree with you.<\/p>\n<p>Sometimes, though, we do not have access to all the data from population that we need, and we have to take a best guess based on a sample. In that case, we probably want to be more conservative in our decision making. That\u2019s what a <strong>t-test<\/strong> is for.<\/p>\n<p>Often, even if we have information regarding the comparison population mean, we do not have access to the population standard deviation (<em>\u03c3<\/em>). In this situation, what we must do is collect a sample reflecting population 1, calculate its mean and standard deviation, and compare that sample mean against a comparison <strong>distribution of means<\/strong> using an estimate (<em>S<\/em><sub>M<\/sub>) of the population (<em>\u03c3<sub>M<\/sub><\/em>). Note the change in symbol (from <em>\u03c3 <\/em>to<em> S)<\/em> to reflect the fact that we are basing our estimate on sample data, rather than using a known population parameter.<\/p>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 39px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-de768ae958fbd82bbc987d55b4a263fb_l3.png\" height=\"39\" width=\"143\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#32;&#92;&#091;&#83;&#94;&#123;&#50;&#125;&#32;&#61;&#32;&#92;&#102;&#114;&#97;&#99;&#123;&#92;&#115;&#117;&#109;&#32;&#40;&#88;&#45;&#77;&#41;&#94;&#123;&#50;&#125;&#125;&#123;&#78;&#45;&#49;&#125;&#92;&#093;&#32;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<p>You may have noticed that this new formula for calculating variance in a sample has on the bottom \u201c<em>N<\/em> minus 1\u201d rather than just \u201c<em>N<\/em>\u201d as our old calculation formula had. Why \u201c<em>N<\/em> minus 1\u201d? This is a correction, based on sample size, that derives from the concept of <strong><a class=\"glossary-term\" aria-haspopup=\"dialog\" aria-describedby=\"definition\" href=\"#term_318_362\">degrees of freedom<\/a><\/strong>. We will return to that concept later. For now, let us consider \u201c<em>N<\/em> minus 1\u201d as a correction based on sample size that has the following deliberate and intentional effects on our calculation. If we have a small sample size, in which we should not have a great deal of confidence, subtracting 1 really affects the calculation. For example, if we had an <em>N<\/em> of 2, subtracting off 1 from that sample size will mean that we divide by 1 instead of 2. That has a drastic impact on the variance calculation, in effect doubling our estimated variance. Of course in the context of hypothesis testing, more variance is bad \u2013 in that we are less likely to be able to make conclusions based on the data.<\/p>\n<p>On the other hand, with a larger sample size, subtracting 1 will have little impact on our calculation\u2026 dividing by 100 or dividing by 99 will have nearly identical results. Think of it this way \u2013 if we have a small sample size, we are punishing ourselves, like a handicap in a golf game. With a larger sample size, we reward ourselves, boosting our chance of a conclusion statistical decision.<\/p>\n<p>Subtracting off 1 is like a flat tax \u2013 a flat income tax of 10% for all Canadians would be harder on those with a low income, who would have even less money to meet basic survival needs, whereas the wealthiest Canadians would still have plenty of money for their needs. That is, of course, why democratic governments typically use progressive tax rates, because those who have more can afford to pay a greater proportion of their income and still have enough to cover their basic costs of living. With hypothesis testing, though, the flat tax is desirable, because it is our intent to penalize the smallest sample sizes the most, given the fact that we can have little confidence in the estimates we derive from them.<\/p>\n<p>The t-test is designed around the concept of <strong>degrees of freedom<\/strong>, which is defined as the number of scores in a given calculation that are free to vary. This is an abstract definition that is difficult to grasp, so it is helpful to consider a concrete example.<\/p>\n<div class=\"textbox textbox--examples\">\n<header class=\"textbox__header\">\n<p class=\"textbox__title\">Degrees of Freedom Practical Example<\/p>\n<\/header>\n<div class=\"textbox__content\">\n<blockquote><p>First, consider the game of baseball.\u00a0 We understand the field-of-play consists of 9 positions.\u00a0 The coach is \u201cfree\u201d to assign any of the 9 players to any of the 9 positions. Once the 8th player is assigned to the 8th position, the 9th player-position is pre-determined, so to speak.\u00a0 In other words, the coach is not \u201cfree\u201d to pick either the last position or the last player.<\/p><\/blockquote>\n<p><em>Source:<\/em> https:\/\/www.isixsigma.com\/ask-dr-mikel-harry\/ask-tools-techniques\/can-you-explain-degrees-freedom-and-provide-example\/<\/p>\n<\/div>\n<\/div>\n<p>The same logic holds when doing a calculation involving all of a sample\u2019s scores. Up until the last score inputted into the formula, the scores are free to vary; but the last score is pre-determined. Thus <strong>degrees of freedom<\/strong> in a <strong>t-test<\/strong> is <em>N<\/em>-1, or 1 less than the total number of scores in the dataset.<\/p>\n<p>In a <strong>t-test<\/strong> we will be using the <strong>t-distributions<\/strong>, rather than a normal distribution, as our comparison distributions. The <strong>t-distributions<\/strong> are a series of distributions, based on the normal distribution, that adjust their shape according to <strong>degrees of freedom<\/strong> (which in turn is based on sample size).<\/p>\n<p>As mentioned previously, when we do not have the population standard deviation we must estimate it from our sample. This involves a greater risk of error. The risk is related to how large our sample is. So, the fewer the scores in our sample, the fewer the degrees of freedom, and the more conservative the distribution we must use \u2013 one that is wider and shorter (more area in the tails).<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-365\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.7-1024x566.jpg\" alt=\"\" width=\"455\" height=\"251\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.7-1024x566.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.7-300x166.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.7-768x424.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.7-1536x849.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.7-65x36.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.7-225x124.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.7-350x193.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.7.jpg 1835w\" sizes=\"auto, (max-width: 455px) 100vw, 455px\" \/><\/p>\n<p>Above you can see examples of how the shape of the t-distributions changes with reduced <strong>degrees of freedom<\/strong>. With <strong>degrees of freedom<\/strong> of 100, the shape of the distribution is very close to the normal curve. If we have very few <strong>degrees of freedom<\/strong>, like 1, then our shading rejection regions, or goal posts, will move farther out into the tails of the distribution, because there is more area under the tails of the curve. In effect, then, with fewer <strong>degrees of freedom<\/strong>, we will find it more difficult to reject the null hypothesis. This is, of course, by design, in order to penalize the hypothesis test if it is based on a very small sample size.<\/p>\n<figure id=\"attachment_366\" aria-describedby=\"caption-attachment-366\" style=\"width: 484px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-366\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.8-1024x609.jpg\" alt=\"\" width=\"484\" height=\"288\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.8-1024x609.jpg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.8-300x178.jpg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.8-768x457.jpg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.8-1536x913.jpg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.8-65x39.jpg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.8-225x134.jpg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.8-350x208.jpg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.8.jpg 1645w\" sizes=\"auto, (max-width: 484px) 100vw, 484px\" \/><figcaption id=\"caption-attachment-366\" class=\"wp-caption-text\"><em>Inspiration for this creative illustration came from one of my former students, Andi.\u00a0Blame for artwork quality is entirely on me.<\/em><\/figcaption><\/figure>\n<p>Above is a less technical visual to see the difference in distribution shape when the sample size is small, versus when the sample size is large. As you can see, when you have a large sample size, the goal posts are very close and easy to reach, but when you have a small sample size, they will really be far away, and this will make it difficult to reject our null hypothesis.<\/p>\n<h4><a id=\"Ttest_return\"><\/a><a href=\"#Ttest\">Concept Practice: Sample size and the t-test<\/a><\/h4>\n<p>&nbsp;<\/p>\n<p>Now we are ready to take a look at how a hypothesis test works based on a single-sample <strong>t-test<\/strong>. As before, the structure of the test is familiar \u2013 we are comparing the means of Population 1 and Population 2.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-369\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.11.png\" alt=\"\" width=\"532\" height=\"325\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.11.png 1001w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.11-300x183.png 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.11-768x468.png 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.11-65x40.png 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.11-225x137.png 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/Fig-5.11-350x213.png 350w\" sizes=\"auto, (max-width: 532px) 100vw, 532px\" \/><\/p>\n<p>As usual, we do not have access to the mean of population 1, just the mean of a sample from that population. Furthermore, unlike in a <strong>Z-test<\/strong>, we do not have access to one piece of information about population 2: the population standard deviation. Therefore, we will form hypotheses the same way, but when we do steps 2 and 4, we will need to use the sample to generate an estimate of the comparison population standard deviation.<\/p>\n<h4><a id=\"ZorT_return\"><\/a><a href=\"#ZorT\">Concept Practice: When to use a Z-test or a t-test<\/a><\/h4>\n<p>&nbsp;<\/p>\n<p>In step 2, we identify the known comparison population\u2019s mean. However, we will need to use the sample to generate an estimate of the comparison population standard deviation. We calculate the sample variance, using the formula with the \u201c<em>N<\/em>-1\u201d correction,<\/p>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 39px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-de768ae958fbd82bbc987d55b4a263fb_l3.png\" height=\"39\" width=\"143\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#32;&#92;&#091;&#83;&#94;&#123;&#50;&#125;&#32;&#61;&#32;&#92;&#102;&#114;&#97;&#99;&#123;&#92;&#115;&#117;&#109;&#32;&#40;&#88;&#45;&#77;&#41;&#94;&#123;&#50;&#125;&#125;&#123;&#78;&#45;&#49;&#125;&#92;&#093;&#32;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<p>and then convert the variance to the <strong>distribution of means<\/strong> by dividing by <em>N<\/em>,<\/p>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 39px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-6635a18895337a42d755b86265f91da3_l3.png\" height=\"39\" width=\"72\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#32;&#92;&#091;&#83;&#94;&#123;&#50;&#125;&#95;&#123;&#77;&#125;&#32;&#61;&#32;&#92;&#102;&#114;&#97;&#99;&#123;&#83;&#32;&#94;&#123;&#50;&#125;&#125;&#123;&#78;&#125;&#92;&#093;&#32;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<p>and finally we square root to get from variance to standard deviation.<\/p>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 32px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-68bfd289ef9453b6d47c1db3733eca96_l3.png\" height=\"32\" width=\"95\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#32;&#92;&#091;&#83;&#95;&#123;&#77;&#125;&#32;&#61;&#32;&#92;&#115;&#113;&#114;&#116;&#123;&#83;&#94;&#123;&#50;&#125;&#95;&#123;&#77;&#125;&#125;&#92;&#093;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<p><em>S<sub>M<\/sub><\/em> is what we will use to describe our comparison distribution.<\/p>\n<p>In step 3, as before we need to find our cutoff\u00a0 sample score based on the significance level and directionality of the test. Now, however, we also need to use a third piece of information: <strong>degrees of freedom<\/strong>. This is found by subtracting one from the sample size. Then we can look in the <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/back-matter\/t-distribution-tables\/\" target=\"_blank\" rel=\"noopener\">t-tables<\/a>, identifying the row of the table with the matching <strong>degrees of freedom<\/strong>, then looking in the appropriate column based on directionality and significance level. Once we identify the t-score that forms the border of the shaded tail, we can map this onto a drawing of the comparison distribution. For visualization purposes, it is fine to sketch the distribution the same way you did the normal distribution.<\/p>\n<h4><a id=\"DF_return\"><\/a><a href=\"#DF\">Concept Practice: Degrees of Freedom<\/a><\/h4>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<p>Now for step 4. The <strong>t-test<\/strong> formula is just like the <strong>Z-test<\/strong> one on top \u2013 sample mean minus comparison population mean. Then you divide by <em>S<sub>M<\/sub><\/em>, the estimate of standard deviation that you calculated back in step 2.<\/p>\n<p class=\"ql-center-displayed-equation\" style=\"line-height: 39px;\"><span class=\"ql-right-eqno\"> &nbsp; <\/span><span class=\"ql-left-eqno\"> &nbsp; <\/span><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/ql-cache\/quicklatex.com-30bb6392ec2914d450cd79778fcd7ad1_l3.png\" height=\"39\" width=\"99\" class=\"ql-img-displayed-equation quicklatex-auto-format\" alt=\"&#32;&#92;&#091;&#116;&#32;&#61;&#32;&#92;&#102;&#114;&#97;&#99;&#123;&#77;&#45;&#92;&#109;&#117;&#32;&#95;&#123;&#77;&#125;&#125;&#123;&#83;&#95;&#123;&#77;&#125;&#125;&#92;&#093;\" title=\"Rendered by QuickLaTeX.com\" \/><\/p>\n<p>If your calculated <strong>t-test<\/strong> score fell into the shaded tail beyond your cutoff score, then you may reject the null hypothesis. In other words, if the sample mean was extreme enough on the comparison distribution of means, reject the null.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter size-large wp-image-377\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/A594CFBD-6310-4E2B-8121-9C3B91C5E9F2-1024x616.jpeg\" alt=\"\" width=\"1024\" height=\"616\" srcset=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/A594CFBD-6310-4E2B-8121-9C3B91C5E9F2-1024x616.jpeg 1024w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/A594CFBD-6310-4E2B-8121-9C3B91C5E9F2-300x180.jpeg 300w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/A594CFBD-6310-4E2B-8121-9C3B91C5E9F2-768x462.jpeg 768w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/A594CFBD-6310-4E2B-8121-9C3B91C5E9F2-1536x923.jpeg 1536w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/A594CFBD-6310-4E2B-8121-9C3B91C5E9F2-65x39.jpeg 65w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/A594CFBD-6310-4E2B-8121-9C3B91C5E9F2-225x135.jpeg 225w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/A594CFBD-6310-4E2B-8121-9C3B91C5E9F2-350x210.jpeg 350w, https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2021\/08\/A594CFBD-6310-4E2B-8121-9C3B91C5E9F2.jpeg 1667w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/p>\n<p>After conducting a hypothesis test, it is helpful to also obtain a report the precise <strong>p-value<\/strong> associated with the result. The <strong>p-value<\/strong> offers the precise probability of obtaining the <strong>t-test<\/strong> result (or more extreme) just by random chance under the comparison distribution. It is found by <a href=\"https:\/\/www.socscistatistics.com\/pvalues\/tdistribution.aspx\" target=\"_blank\" rel=\"noopener\">calculating<\/a> the area under the comparison distribution in the tail beyond the observed <strong>t-test<\/strong> score calculated in step 4 (or in both tails, if it was a two-tailed test).<\/p>\n<p>&nbsp;<\/p>\n<h4><a id=\"tTest_return\"><\/a><a href=\"#tTest\">Concept Practice: Conduct a t-test<\/a><\/h4>\n<h4><a id=\"Facts_return\"><\/a><a href=\"#Facts\">Concept Practice: Facts about Z- and t-tests<\/a><\/h4>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<h1>Chapter Summary<\/h1>\n<p>In this chapter we learned about the two types of logical errors we can make in hypothesis testing: <strong>Type I<\/strong> and <strong>Type II errors<\/strong>. We saw that <strong>Type I error<\/strong> is considered more serious in statistics, because it represents going out on limb, making a strong conclusion, when that is in fact the wrong decision. For that reason, <strong>Type I error<\/strong> risk is strictly controlled in inferential statistics by setting a particular significance level (<em><strong>\u03b1<\/strong><\/em>). We also introduced our first real inferential statistical tests: the single-sample <strong>Z-test<\/strong> and the single-sample <strong>t-test<\/strong>. In each of these tests, we are comparing the means of two populations, using a sample to estimate the mean of the research population. A <strong>Z-test<\/strong> is used when we know the standard deviation of the comparison population (<em><strong>\u03c3<\/strong><\/em>); a <strong>t-test<\/strong> is used when we do not have that information and must estimate the standard deviation from the sample (<strong><em>S<\/em><\/strong>).<\/p>\n<p>Key terms:<\/p>\n<table class=\"no-lines\" style=\"border-collapse: collapse;width: 100%\">\n<tbody>\n<tr>\n<td style=\"width: 33.3333%\"><strong>Type I error<\/strong><\/td>\n<td style=\"width: 33.3333%\"><strong>distribution of means<\/strong><\/td>\n<td style=\"width: 33.3333%\"><strong>p-value<\/strong><\/td>\n<\/tr>\n<tr>\n<td style=\"width: 33.3333%\"><strong>Type II error<\/strong><\/td>\n<td style=\"width: 33.3333%\"><strong>Z-test<\/strong><\/td>\n<td style=\"width: 33.3333%\"><strong>t-test<\/strong><\/td>\n<\/tr>\n<tr>\n<td style=\"width: 33.3333%\"><em><strong>\u03b1<\/strong><\/em><\/td>\n<td style=\"width: 33.3333%\"><strong>standard error of the mean<\/strong><\/td>\n<td style=\"width: 33.3333%\"><strong>degrees of freedom<\/strong><\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<h1>Concept Practice<\/h1>\n<p><a id=\"ErrorTypes\"><\/a><\/p>\n<div id=\"h5p-31\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-31\" class=\"h5p-iframe\" data-content-id=\"31\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5a.02. Type1 error\"><\/iframe><\/div>\n<\/div>\n<div id=\"h5p-32\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-32\" class=\"h5p-iframe\" data-content-id=\"32\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5a.03. Type1 and Type 2 errors.\"><\/iframe><\/div>\n<\/div>\n<div id=\"h5p-33\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-33\" class=\"h5p-iframe\" data-content-id=\"33\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5a.04. Type1 and type2 errors.\"><\/iframe><\/div>\n<\/div>\n<div id=\"h5p-34\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-34\" class=\"h5p-iframe\" data-content-id=\"34\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5a.05. Type1 and type2 errors.\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#ErrorTypes_return\">text<\/a><\/h6>\n<p><a id=\"DistMeans\"><\/a><\/p>\n<div id=\"h5p-37\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-37\" class=\"h5p-iframe\" data-content-id=\"37\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5a.08. Distribution of means\"><\/iframe><\/div>\n<\/div>\n<div id=\"h5p-38\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-38\" class=\"h5p-iframe\" data-content-id=\"38\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5a.09. continued. Dstribution of means.\"><\/iframe><\/div>\n<\/div>\n<div id=\"h5p-39\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-39\" class=\"h5p-iframe\" data-content-id=\"39\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5a.10. continued. Distribution of means.\"><\/iframe><\/div>\n<\/div>\n<div id=\"h5p-40\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-40\" class=\"h5p-iframe\" data-content-id=\"40\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5a.11. continued. Distribution of means.\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#DistMeans_return\">text<\/a><\/h6>\n<p><a id=\"Ztest\"><\/a><\/p>\n<div id=\"h5p-41\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-41\" class=\"h5p-iframe\" data-content-id=\"41\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5a.12. Automatic braking system hypothesis testing\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#Ztest_return\">text<\/a><\/h6>\n<p><a id=\"Ttest\"><\/a><\/p>\n<div id=\"h5p-43\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-43\" class=\"h5p-iframe\" data-content-id=\"43\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5b.03. Degree of freedom and sample size.\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#Ttest_return\">text<\/a><\/h6>\n<p><a id=\"ZorT\"><\/a><\/p>\n<div id=\"h5p-44\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-44\" class=\"h5p-iframe\" data-content-id=\"44\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5b.01. When to use t-test.\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#ZorT_return\">text<\/a><\/h6>\n<p><a id=\"DF\"><\/a><\/p>\n<div id=\"h5p-42\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-42\" class=\"h5p-iframe\" data-content-id=\"42\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5b.02. Degrees of freedom\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#DF_return\">text<\/a><\/h6>\n<p><a id=\"tTest\"><\/a><\/p>\n<div id=\"h5p-46\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-46\" class=\"h5p-iframe\" data-content-id=\"46\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5b.05. T-test hypothesis testing for rotating leader example\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#tTest_return\">text<\/a><\/h6>\n<p><a id=\"Facts\"><\/a><\/p>\n<div id=\"h5p-45\">\n<div class=\"h5p-iframe-wrapper\"><iframe id=\"h5p-iframe-45\" class=\"h5p-iframe\" data-content-id=\"45\" style=\"height:1px\" src=\"about:blank\" frameBorder=\"0\" scrolling=\"no\" title=\"Practice 5b.04. t-test basic concepts.\"><\/iframe><\/div>\n<\/div>\n<h6 style=\"text-align: right\">Return to <a href=\"#Facts_return\">text<\/a><\/h6>\n<p>Return to <a href=\"#5a\">5a. Single-sample Z test<\/a><\/p>\n<h6>Try interactive <a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/?post_type=chapter&amp;p=889&amp;preview=true\">Worksheet 5a<\/a> or download Worksheet 5a<\/h6>\n<p>Return to <a href=\"#5b\">5b. Single-sample t test<\/a><\/p>\n<h6>Try interactive Worksheet 5b\u00a0or download Worksheet 5b<\/h6>\n<p>&nbsp;<\/p>\n<p><a id=\"video5a\"><\/a>video 5a1<\/p>\n<div style=\"width: 1440px;\" class=\"wp-video\"><video class=\"wp-video-shortcode\" id=\"video-318-1\" width=\"1440\" height=\"1080\" preload=\"metadata\" controls=\"controls\"><source type=\"video\/mp4\" src=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/08\/Chapter-5a1.mp4?_=1\" \/><a href=\"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/08\/Chapter-5a1.mp4\">https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-content\/uploads\/sites\/1469\/2022\/08\/Chapter-5a1.mp4<\/a><\/video><\/div>\n<p>video 5a2<\/p>\n<p>&nbsp;<\/p>\n<p><a id=\"video5b\"><\/a>video 5b1<\/p>\n<div class=\"glossary\"><span class=\"screen-reader-text\" id=\"definition\">definition<\/span><template id=\"term_318_320\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_318_320\"><div tabindex=\"-1\"><p>if we made the decision to reject the null hypothesis when it is true <\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_318_321\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_318_321\"><div tabindex=\"-1\"><p>if we made the decision to not reject the null hypothesis but the research hypothesis is true<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_318_580\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_318_580\"><div tabindex=\"-1\"><p>the probability of making a Type I error; used as shorthand for significance level<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_318_328\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_318_328\"><div tabindex=\"-1\"><p>also called a sampling distribution, is the distribution of many sample means drawn from the population of individual scores<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_318_336\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_318_336\"><div tabindex=\"-1\"><p>statistical hypothesis test suitable for comparing the means of two populations, when the comparison population mean and standard deviation are known<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_318_334\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_318_334\"><div tabindex=\"-1\"><p>standard deviation of the distribution of means<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_318_352\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_318_352\"><div tabindex=\"-1\"><p>the probability of the observed sample score or more extreme occurring at random under the comparison distribution<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_318_358\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_318_358\"><div tabindex=\"-1\"><p>statistical test to test the differences between two population means. Suitable for single sample design when standard deviation is unknown, or in two-sample designs.<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><template id=\"term_318_362\"><div class=\"glossary__definition\" role=\"dialog\" data-id=\"term_318_362\"><div tabindex=\"-1\"><p>the number of scores in a given calculation that are free to vary.<\/p>\n<\/div><button><span aria-hidden=\"true\">&times;<\/span><span class=\"screen-reader-text\">Close definition<\/span><\/button><\/div><\/template><\/div>","protected":false},"author":1394,"menu_order":5,"template":"","meta":{"pb_show_title":"on","pb_short_title":"","pb_subtitle":"","pb_authors":[],"pb_section_license":""},"chapter-type":[48],"contributor":[],"license":[],"class_list":["post-318","chapter","type-chapter","status-publish","hentry","chapter-type-numberless"],"part":3,"_links":{"self":[{"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/chapters\/318","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/chapters"}],"about":[{"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/wp\/v2\/types\/chapter"}],"author":[{"embeddable":true,"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/wp\/v2\/users\/1394"}],"version-history":[{"count":25,"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/chapters\/318\/revisions"}],"predecessor-version":[{"id":1236,"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/chapters\/318\/revisions\/1236"}],"part":[{"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/parts\/3"}],"metadata":[{"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/chapters\/318\/metadata\/"}],"wp:attachment":[{"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/wp\/v2\/media?parent=318"}],"wp:term":[{"taxonomy":"chapter-type","embeddable":true,"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/pressbooks\/v2\/chapter-type?post=318"},{"taxonomy":"contributor","embeddable":true,"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/wp\/v2\/contributor?post=318"},{"taxonomy":"license","embeddable":true,"href":"https:\/\/pressbooks.bccampus.ca\/statspsych\/wp-json\/wp\/v2\/license?post=318"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}