Key Terms List
 α

the probability of making a Type I error; used as shorthand for significance level
 β

the probability of making a Type II error; the antithesis of power
 Σ

in summation notation, a symbol that denotes “taking the sum” of a series of numbers
 Analysis of Variance

also called ANOVA, a system of data analysis that is very flexible and adaptable to a variety of research designs. It is based on a statistical concept called the general linear model and involves the technique of partitioning variance.
 bimodal

a descriptor of a distribution indicating that there are two peaks, or two collections of scores
 Bonferroni correction

adjustment to avoid inflation of experimentwise risk of Type I error, by dividing significance level by the number of planned contrasts to be conducted
 central limit theorem

mathematical theorem that proposes the following: as long as we take a decentsized sample, if we took many samples (10,000) of large enough size (30+) and took the mean each time, the distribution of those means will approach a normal distribution, even if the scores from each sample are not normally distributed
 central tendency

a statistical measure that defines the centre of a distribution with a single score
 Cohen’s d

a measure of effect size commonly used to quantify the difference between two population means; 0.2 is small, 0.5 is medium, and 0.8 is considered large
 confidence intervals

an approach to inferential statistics that serves as an alternative to hypothesis testing. A statement of where a research population mean should lie with a particular probability.
 correlation

statistical analysis of the direction and strength of the relationships between two numerical variables
 covariance

the variability that two numeric variables have in common
 cutoff sample score

critical value that serves as a decision criterion in hypothesis testing
 degrees of freedom

the number of scores in a given calculation that are free to vary.
 dependent means ttest

a test for statistical significance when comparing mean difference scores to zero in repeated measures or matched pairs designs
 dependent variables

a variable you measure to detect a difference/change as a result of the manipulation  most often it is numeric
 descriptive

ways to summarize or organize data from a research study – essentially allowing us to describe what the data are
 directional hypothesis

a research prediction that the research population mean will be “greater than” or "less than" the comparison population mean
 distribution of means

also called a sampling distribution, is the distribution of many sample means drawn from the population of individual scores
 do not reject the null hypothesis

a decision in hypothesis testing that is inconclusive because the sample score is less extreme than the cutoff score
 effect size

a measure of how well a statistical model explains variability, apart from statistical significance, e.g. how big a difference between means is, or just how much variability a regression model explains
 experimentwise alpha level

the problem of accumulating risk of Type I error with multiple statistical tests on the same data
 factor

in ANOVA, a grouping variable used to account for variance among scores; in an experiment a factor is an independent variable
 frequency tables

a way to summarize a dataset in table form, to organize the data and make it easy to get an overview of the dataset quickly
 general linear model

an extension of the statistical technique linear regression that is adaptable to various combinations of independent (nominal) and dependent (numeric) variables
 grouped frequency table

a frequency table that defines ranges of values in the first column, and reports the frequency of scores that fall within each range
 histogram

a graph for summarizing numeric data that essentially is a frequency table that has been turned on its side, with the added benefit of a visual representation of the frequency as the height of the bars in the graph, rather than just a number
 homoscedasticity assumption

independent means ttests require the assumption that the two populations we are comparing have the same variance
 hypothesis testing

a formal decision making procedure often used in inferential statistics
 independent means ttest

a statistical test used in hypothesis tests comparing the means of two independent samples, created by random assignment of individuals to experimental and control groups
 independent variables

a variable you manipulate  most often it is categorical, or nominal
 inferential

analytical tools that allow us to draw conclusions based on data from a research study  essentially allowing us to make a statement about what the data mean
 interaction

the degree to which the contribution of one factor to explaining variability in the data depends on the other factor; the synergy among factors in explaining variance
 left skewed

a descriptor of a distribution that indicates asymmetry, specifically with a low frequency tail leading off to the left
 levels

the individual conditions or values that make up a factor, a nominal variable that forms the groups in analysis of variance
 levels of statistical significance

the probability level that we are willing to accept as a risk that the score from our research sample might occur by random chance within the comparison distribution. By convention, it is set to one of three levels: 10%, 5%, or 1%.
 M

the symbol for the mean (average) of scores in a sample
 main effect

the degree to which one of the factors explains variability in the data when taken on its own, independent of the other factor
 matched pairs

a research design for which a dependent means ttest may be used to test for a hypothesis test; in this design two separate samples are used, but each individual in a sample is matched onetoone with an individual in the other sample, most often matching participants on a a possible confounding variable as a way to control for the effects of that variable
 mean

the same thing as an average: you add up all the numbers, then divide by how many numbers there were. Conceptually we can think of the mean as the balancing point for the distribution.
 median

the midpoint of the scores after placing them in order. The median is a countingbased measure: the point at which half of the scores fall above and half of the scores fall below.
 mode

the score(s) that occur(s) most often in the dataset
 N

the symbol for the number of scores in a sample
 nominal

variables that label or categorize something, and any numbers used to measure these variables are arbitrary and do not indicate quantity or size
 nondirectional hypothesis

a research prediction that the research population mean will be “different from" the comparison population mean, but allows for the possibility that the research population mean may be either greater than or less than the comparison population mean
 normal curve

a theoretical distribution, sometimes called a Z distribution, has a very distinct set of properties that make it a useful model for data analysis (e.g. 21434% area rule)
 normal curve assumption

parametric tests like the ttest and Ztest require the assumption that the distribution of means for any given population is normally distributed
 null hypothesis

the prediction that the population from which sample came is not different from the comparison population
 numeric

variables for which numbers are actually meaningful  they indicate the size or amount of something
 onetailed test

a hypothesis test in which there is only one cutoff sample score on either the lower or the upper end of the comparison distribution
 pvalue

the probability of the observed sample score or more extreme occurring at random under the comparison distribution
 participant variables

variables used like independent variables in (quasi)experimental research designs, but which cannot be manipulated or assigned randomly to participants, and as such must not generate causeeffect conclusions
 partitioning variance

the allocation of variability among scores in numeric data into different buckets, like treatment effects vs. error, or betweengroups vs. withingroups variance
 percentile

the score at which a given percentage of scores in the normal distribution fall beneath
 planned contrasts

statistical tests of pairwise comparisons among groups, used to follow up on a significant ANOVA result, when researchers know in advance which groups they expect to differ
 population

all possible individuals or scores about which we would ideally draw conclusions
 posthoc tests

statistical tests of pairwise comparisons among groups, used to follow up on a significant ANOVA result, when researchers do not know in advance which groups they expect to differ and wish to test all possible combinations
 power

the probability of rejecting the null hypothesis (i.e. finding statistical significance) if the research hypothesis is in fact true. Depends on effect size and sample size.
 probability

in a situation where several different outcomes are possible, the probability of any specific outcome is a fraction or proportion of all possible outcomes
 r

correlation coefficient that describes the strength and direction of the relationship between two numeric variables. Can be between 1 and 0 and between 0 and +1.
 r squared

proportion of variability in one variable that can be explained by the relationship with the other variable. Can be between 0 and 1.
 regression

a statistical model that allows for prediction based on a trend line that “best fits” the data points that we have collected. Mathematically, a regression line is one that minimizes the squared deviations (i.e. error) of each point from the line.
 reject the null hypothesis

a decision in hypothesis testing that concludes statistical significance because the sample score is more extreme than the cutoff score
 repeated measures

also known as withinsubjects designs or pretest posttest design, in which the experiment involves obtaining two separate scores for each individual in a single sample. The same participants are used in all treatment conditions.
 research hypothesis

prediction that the population from which the research sample came is different from the comparison population
 right skewed

a descriptor of a distribution that indicates asymmetry, specifically with a low frequency tail leading off to the right
 sample

the individuals or scores about which we are actually drawing conclusions
 Scheffe’s correction

in posthoc analyses, an adjustment to correct for inflated experimentwise risk of Type I error, by dividing the F value by the overall degrees of freedom between from the original overall ANOVA analysis
 score

a particular individual’s value on the variable
 standard deviation

a common measure of variability in numeric data. The average distance of a scores from the mean.
 standard error of the mean

standard deviation of the distribution of means
 statistically significant

the conclusion from a hypothesis test that probability of the observed result occurring randomly within the comparison distribution is less than the significance level
 Sum of Squares

the sum of squared deviations, or differences, between scores and the mean in a numeric dataset
 tdistributions

a series of distributions, based on the normal distribution, that adjust their shape according to degrees of freedom (which in turn is based on sample size)
 ttest

statistical test to test the differences between two population means. Suitable for single sample design when standard deviation is unknown, or in twosample designs.
 twotailed test

a hypothesis test in which there are two cutoff sample scores, one on either end of the comparison distribution
 Type I error

if we made the decision to reject the null hypothesis when it is true
 Type II error

if we made the decision to not reject the null hypothesis but the research hypothesis is true
 μM

population mean for the distribution of means
 unimodal

a descriptor of a distribution indicating that there is one peak, or a single collection of scores
 value

any possible number or category that a variable could take on
 variable

a quality or a quantity that is different for different individuals
 variance

a common measure of variability in numeric data. The average squared distance of scores from the mean.
 Zscores

standard scores that allow us to transform scores in any numeric dataset, using any scale, into a standard metric
 Ztest

statistical hypothesis test suitable for comparing the means of two populations, when the comparison population mean and standard deviation are known