Tables and Formulas for Sullivan, Statistics Informed Decisions

Tables and Formulas for Sullivan, Statistics: Informed Decisions Using Data

Chapter 2: Organizing and Summarizing Data

Relative frequency = frequencysum of all frequenciesClass midpoint: The sum of consecutive lower class limits divided by 2.

Chapter 3: Numerically Summarizing Data

Population Mean: μ= ΣxiNSample Mean: x = ΣxinRange = Largest Data Value – Smallest Data Value

Population Standard Deviation: σ = Σ(xi- μ)2NSample Standard Deviation: s = Σ(xi- x)2n-1Population Variance: σ2Sample Variance: s2

Empirical Rule: If the shape of the distribution is bell-shaped, then

Approximately 68% of the data will lie within 1 standard deviation of the mean

Approximately 95% of the data will lie within 2 standard deviations of the mean

Approximately 99.7% of the data will lie within 3 standard deviations of the mean

Population z-score: z = x- μσSample z-score: z = x- xsInterquartile Range: IQR = Q3 – Q1Lower Fences = Q1-1.5(IQR)

Upper Fence = Q3+1.5(IQR)Five-Number Summary

Minimum, Q1, M, Q3, MaximumChapter 4: Describing the Relation between Two Variables

Correlation Coefficient: -1 ≤ r ≤ 1

The equation of the least-squares regression line is y = b1x +b0,

where y is the predicted value, b1 is the slope, and b0 is the y-intercept.

Residual = observed y – predicted y

= y – yThe coefficient of determination, R2, measures the proportion of total variation in the response variable that is explained by the least-squares regression line.

Chapter 5: Probability

Empirical Probability

P(E) = frequency of Enumber of trials of experimentClassical Probability

P(E) = number of ways that E can occurnumber of possible outcomes = N(E)N(S)Addition Rule for Disjoint Events

P(E or F) = P(E) + P(F)

General Addition Rule

P(E or F) = P(E) + P(F) – P(E and F)

Compliment Rule

P(EC) = 1– P(E)

Multiplication Rule for Independent Events

P(E and F) = P(E)⋅P(F)

Conditional Probability Rule

P (F|E) = P(E and F)P(E)= N(E and F)N(E)General Multiplication Rule

P (E and F) = P(E) ⋅ P(F|E)

Chapter 6: Discrete Probability Distributions

Mean (Expected Value) of a Discrete Random Variable

μX = Σx ⋅ P(x)

Standard Deviation of a Discrete Random Variable

σX = Σ (x- μ)2 ⋅P(x)

Binomial Probability Distribution Function

13970002032000P(x) = n C x px(1-p)n-xMean and Standard Deviation of a Binomial Random Variable

μX=npσX = np(1-p)

Chapter 7: The Normal Distribution

Standardizing a Normal Random Variable

z = x – μσ

Finding the Score: x = μ+zσChapter 8: Sampling Distributions

Mean and Standard Deviation of the Sampling Distribution of xμx = μ andσx = σnSample Proportion: p = xnMean and Standard Deviation of the Sampling Distribution of pμp = p and σp = p(1-p)nChapter 9: Estimating the Value of a Parameter

Confidence Intervals

A (1 – α)⋅100% confidence interval about p is p ± zα/2⋅p(1 – p)nA (1 – α) ⋅100% confidence interval about μ is x ± tα/2⋅snNote: tα/2 is computed using n – 1 degrees of freedom.

Sample Size

To estimate the population proportion with a margin of error E at a (1 – α)⋅ 100% level of confidence:

n = p (1 – p) zα/2Ε2 rounded up to the next integer, where p is a prior estimate of the population proportion, or n = 0.25 zα/2Ε2rounded up to the next integer when no prior estimate of p is available.

To estimate the population mean with a margin of error E at a (1 – α) ⋅ 100% level of confidence:

n = zα/2⋅ sΕ2 rounded up to the next integer.

Chapter 10: Hypothesis Tests Regarding a Parameter

Test Statistics

z0 = p- p0p0 (1- p0)nt0= x – μ0 snChapter 11: Inferences on Two Samples

Test Statistic Comparing Two Population Proportions (Independent Samples)

z0 = p1- p2-(p1- p2)σ(p1-p2) where p = x1+ x2n1+ n2Confidence Interval for the Difference of Two Proportions (Independent Samples)

(p1- p2) ± zα/2 ⋅ σp1- p2Test Statistic for Matched-Pairs Data

t0= d- μdsdnwhere d is the mean and sd is the standard deviation of the differenced data.

Confidence Interval for Matched-Pairs Data

d ± tα2 ⋅ sdnNote: tα2 is found using n-1 degrees of freedom.

Test Statistic Comparing Two Means (Independent Sampling)

t0 = x1-x2-(μ1-μ2)s12n1+ s22n2Confidence Interval for the Difference of Two Means (Independent Samples)

(x1 – x2) ± tα2s12n1+ s22n2Note: tα2 is found using the smaller of n1-1 or n2-1 degrees of freedom.

Chapter 12: Additional Inferential Procedures

Chi-Square Procedures

Expected Counts (when testing for goodness of fit)

Εi = μi = npi for i = 1, 2, …, k

Expected frequencies (when testing for independence or homogeneity of proportions)

Expected frequency = (row total)(column total)table totalChi-Square Test Statistic

02 = (observed-expected)2expected = (Oi – Ei)2 Ei i = 1, 2, …, k

All Ei ≥ 1 and no more than 20% less than 5.

Tables:

Table II

Critical Values (CV) for Correlation Coefficient

n CV

3 0.997

4 0.950

5 0.878

6 0.811

7 0.754

8 0.707

9 0.666

10 0.632

11 0.602

12 0.576

13 0.553

14 0.532

15 0.514

16 0.497

17 0.482

18 0.468

19 0.456

20 0.444

21 0.433

22 0.423

23 0.413

24 0.404

25 0.396

26 0.388

27 0.381

28 0.374

29 0.367

30 0.361

Table VI

Critical Values for Normal Probability Plots

Sample Size, n Critical Value

5 0.880

6 0.888

7 0.898

8 0.906

9 0.912

10 0.918

11 0.923

12 0.928

13 0.932

14 0.935

15 0.939

16 0.941

17 0.944

18 0.946

19 0.949

20 0.951

21 0.952

22 0.954

23 0.956

24 0.957

25 0.959

30 0.960