PSYC 60 Quiz 2 (Ch.7-11) PDF
Document Details

Uploaded by SeamlessMoonstone8761
UC San Diego
Tags
Summary
This document appears to be a quiz, covering material from chapters 7-11 on topics such as sampling distributions, hypothesis testing, and t-tests. Relevant concepts include standard error, null and alternative hypotheses, and calculating test statistics.
Full Transcript
Ch. 7 Sampling Distributions Sampling distribution: a probability distribution of all statistics (e.g. all sample means) of a given sample size from a population. ○ The sampling distribution of the mean is the most common sampling distribution Mean...
Ch. 7 Sampling Distributions Sampling distribution: a probability distribution of all statistics (e.g. all sample means) of a given sample size from a population. ○ The sampling distribution of the mean is the most common sampling distribution Mean of a sampling distribution of the mean: μM The variance of the sampling distribution of the mean is denoted: σ²M Standard error: the standard deviation of a sampling distribution Measure of sampling error ○ Sampling error: the difference between a statistic and the parameter It’s the variability of a statistic from sample to sample due to chance; not mistakes in sampling. The bigger the standard error the bigger the sampling error.** ○ The standard error changes with sample size and the population standard deviation. A smaller population standard deviation equals a smaller standard error. The less that population scores deviate from the mean, the less that sample scores can deviate from the population mean. Increasing the sample size will decrease standard error (decrease sampling error) because of the law of large numbers. ○ Law of large numbers: the law that increasing sample size decreases standard error. The larger the sample size, the more likely the sample’s statistics will resemble the corresponding parameters. Hence, larger samples are associated with more accurate estimates of parameters. Central limit theorem: a mathematical statement regarding the nature of sampling distributions ○ The theorem is that given a population with a mean μ and a variance σ², the sampling distribution of the mean will: Have a mean equal to μ Have a variance equal to σ²/N (N = the sample size, NOT the number of samples) Approach the normal distribution as N increases, regardless of the shape of the population. ○ Ex: ○ The central Limit theorem is important because it allows for inferential statistics, ex, allows scientists to draw conclusions about populations from samples. ** Ch. 8 Hypothesis Testing Null Hypothesis Significance Testing Hypothesis: a statement, e.g., the value of a parameter, or a tentative explanation of a phenomenon ○ Example statement: the population mean final grade for undergraduate statistics is 80% ○ Example tentative explanation: statistics is challenging to learn because the concepts are often not intuitive. Null hypothesis (H0): a statement about a parameter,e.g., a population mean, that is ASSUMED to be true. ○ It's usually a hypothesis of no difference or no relationship between variables (innocent until proven guilty) ○ E.g., a null hypothesis could be that there is NO relationship between studying and test performance. ○ The alternative to a null hypothesis is the alternative hypothesis. Alternative/research hypothesis (H1): a statement that directly contradicts a null hypothesis. ○ It’s the hypothesis that usually the researcher BELIEVES TO BE TRUE ○ E.g., an alternative hypothesis could be that there IS a relationship between studying and test performance. ○ H0 and H1 must encompass ALL possibilities for a population, e.g., studying either is or is not related to test performance. ○ The researcher’s aim (in almost all cases) is to reject the null hypothesis Null hypothesis (H0) Alternative/research hypothesis (H1) There is NOT a relationship between There IS a relationship between studying and test performance. studying and test performance. Null hypothesis significance testing (NHST): the evaluation of statistics to estimate parameters. ○ NHST calculates the probability of obtaining a statistic if the hypothesis regarding the parameter is true. ○ STEPS: (1) State the hypothesis, i.e., state the null hypothesis (H0) and research hypothesis (H1) (2) Set the criteria for rejecting/not rejecting the null hypothesis (3) Compute the test statistic, i.e., calculate the probability of obtaining a statistic if H0 is true. (4) Make a decision; ex: reject or do not reject the null hypothesis. One- and Two-Tailed Tests Hypotheses are often evaluated with one- or two- tailed tests. One-tailed (directional) test Two-tailed (nondirectional test) A hypothesis test in which the value in A hypothesis test in which the value in H1 is states as > OR < the value in H0 H1 is stated as ≠ the value in H0 Ex: the population mean final grade Ex: the population mean final grade for undergraduate statistics is > 80% for undergraduate statistics ≠ 80% Ex: the population mean final grade for undergraduate statistics is < 80% Alpha (α): the significance level for a hypothesis test ○ It’s usually set at 0.05 (α = 0.05) in behavioral research ○ 5% of the distribution will serve as the rejection region. Calculating a Test Statistic Test statistic: a formula to determine the probability of obtaining a statistic if the null hypothesis is true. ○ If a test statistic is GREATER than a critical value, then REJECT the null hypothesis. Critical value: a cutoff value ○ Ex: the critical values on slide 18 = ±1.96. Thus, a test statistic that is < -1.96 or > +1.96 would indicate to REJECT H0 One-sample z test: an inferential statistic that uses z scores to determine if a sample mean is significantly different from a population mean ○ ○ ○ P value P value: the probability of obtaining a test statistic that is at least as extreme as the observed result if H0 is true. ○ P values determine whether H0 should be rejected. ○ If p < α, REJECT H0 ○ If p > α, DO NOT REJECT H0 Type I and Type II Errors Type I error: the error of rejecting H0 when H0 is true (false positive) ○ The probability of making a Type I error = α Type II error: the error of NOT REJECTING H0 when H0 is false ○ The probability of making a Type II error = β Power: the probability of correctly rejecting a false H0 (THIS IS GOOD) ○ The probability = 1- β Increasing the Type I error rate decreases the type II error rate, and vice versa, hence we don’t set either to zero. Ch. 9 One-sample T test Definition and Examples T test: an inferential statistic for determining if there is a significance difference between two means. ○ T tests include one sample t tests, dependent-samples t tests, and independent-samples t tests One-sample t test: an inferential statistic for determining if there is a significant difference between a sample mean and population mean. Example questions include for ONE-SAMPLE T TEST: ○ Is there a difference in graduate record examination (GRE) scores between a sample of Californians and the population of Americans? ○ Is there a difference in the personality trait agreeableness between a sample of 20-year-olds and the general population. ○ Is there a difference in working memory capacity between a sample of STEM students and the general population? ○ The formula for the standard error of the mean is the denominator of the one-sample z test: (σ/√N) ○ The denominator of the one-sample t test (SD/√N) provides an estimate of the standard error of the mean. Z tests use the standard normal distribution; t tests use the t distribution (df increase with sample size) ○ Assumptions Two assumptions must be met for one-sample t tests to be valid: ○ (1) The DV is measured with a ratio or interval scale. ○ (2) The sample is drawn from a normally distributed population. Normality is met: ○ Skewness coefficient is between ±1 (SYMMETRY) Positive coefficient (i.e.,> 0) for skew indicates positive skew. Negative coefficient (i.e., < 0) for skew indicates negative skew. Coefficient of 0 for skew indicates a symmetrical distribution. ○ Kurtosis coefficient is between ±2 (POINTEDNESS) Kurtosis of the standard normal distribution is 3 (mathematically), but set to 0 in SPSS Positive coefficients for kurtosis indicate a more pointed distribution, i.e., leptokurtic Negative coefficients for kurtosis indicate a less pointed distribution, i.e., platykurtic. Standard normal distribution is mesokurtic Calculating t Effect Size Effect size: a measure of the magnitude of a relationship between variables ○ Many effects are statistically significant with large enough samples but not necessarily meaningful or important. ○ Effect size informs if a significant result is meaningful or important. Cohen’s d: a measure of effect size in terms of the number of standard deviations that mean differ from each other. (estimated) ○ The most common measure of effect size for t tests ○ It’s estimated because statistics are used instead of parameters ○ Example: Negatives could happen when sample mean is smaller than population mean! ○ Interpreting Cohen’s D value: ○ In other words using the previous example except with N = 25… Confidence Interval Confidence interval (CI): a range of values that likely contains an unknown parameter. ○ Calculating a confidence interval provides a way to: Estimate μ (in our case, μ for STEM students) Infer statistical significance Estimate effect size ○ Example: Interpretation: At a 95% confidence interval, the estimated population mean of STEM students’ working memory capacity is between 4.09 and 4.91 Factors that Influence Power Power: the probability of correctly rejecting a false H0 (THIS IS GOOD) ○ The probability = 1- β Factors that influence power Meaning Effect size Effect increase, power increase - As the difference between M and μ increases, power increases Sample size N increase, decrease sampling power and increase power - Imagine a sample size increasing to the point it includes the entire population. Sampling error would decrease 0. M would equal μ Alpha (α) Increasing α, will increase power Beta (β) Increasing β, will decrease power - 1-β = power; therefore, increasing beta decreases power. Standard Deviation (SD) As SD increases, power decreases - Populations with more dispersion of scores produce samples with more dispersion. - More dispersion equals more sampling error, and less certainty that samples are representative of populations. Ch. 10 Independent-Samples t Test Definition and Examples Independent-samples t test: an inferential statistic for determining if there is a significant difference between two independent/unrelated samples. ○ Independent-samples t tests are a type of between-subject designs. Between-subjects design: a research method in which participants are assigned to only one of multiple conditions, and then scores are obtained and compared across conditions. Examples question for INDEPENDENT-SAMPLES T TESTS: ○ Is there a difference in graduate record examination (GRE) scores between a sample of Californians and a sample of Delawareans? ○ Is there a difference in the personality trait agreeableness between a sample of 20-year olds and a sample of 60-years-olds? ○ Is there a difference in working memory capacity between a sample of STEM students and a sample of art students? Assumptions Three assumption must be met for independent-samples t tests to be valid: ○ (1) The DV is measured with a ratio or interval scale. ○ (2) The samples are drawn from normally distributed populations ○ (3) The samples are drawn from populations with equal (homogenous) variances. Calculating T **The denominator is the estimated standard error for the difference between means.** Effect Size Confidence Interval For Independent samples, the CI provides a way to: ○ Estimate the difference between μ1 and μ2 ○ Infer statistical significance (DON'T HAVE TO DO ON TEST) ○ Estimate effect size (DON'T HAVE TO DO ON TEST) Ch. 11 Dependent-Samples t Test Definition and Examples Dependent-samples T test: an inferential statistic for determining if there is a significant difference between two dependent/related samples. ○ Dependent-samples t tests are a type of within-subjects design. Within-subjects design: a research method in which participants are assigned to multiple conditions, and then scores are obtained and compared across conditions. Example questions for DEPENDENT-SAMPLES T TESTS: ○ Is there a difference between in a sample of individuals’ graduate record examination (GRE) scores from before to after a GRE prep course? ○ Is there a difference in a sample of 20-year-olds’ agreeableness scores from two different personality measures? ○ Is there a difference in a sample of STEM students’ working memory capacity from before to after caffeine intake? Assumptions Two assumptions must be met for dependent-sample t tests to be valid: ○ (1) the DV is measured with a ratio or interval scale ○ (2) The samples are drawn from normally distributed populations Calculating T ○ Subscript “d” is the difference of the “before” and “after” scores ○ The denominator is the estimated standard error for the mean difference scores. Effect Size Confidence Interval For dependent samples, the CI provides a way to: ○ Estimate μD ○ Infer statistical significance (DON'T HAVE TO DO ON TEST) ○ Estimate effect size (DON'T HAVE TO DO ON TEST) Advantages of Within-Subjects Design The advantages of within-subjects (w/n-Ss) designs vs. between-subjects (b/n-Ss) designs are that the former increase: ○ (1) power, by decreasing standard error by: (a) eliminating person-to-person variability Ex: ALTERNATIVELY SUPPOSE: (b) increasing control of extraneous variables Ex: ALTERNATIVELY SUPPOSE: ○ (2) practicality Recruiting participants require time, energy, and often money For the same amount of power as a b/n-Ss design, w/n-Ss design requires fewer participants, and thus costs less time, energy, and money.