Warne 1st edition Full Test Bank - Statistics for the Social Sciences 1st Edition Test Bank by Russell T. Warne. DOCX document preview.
Test Bank
Notes
The questions in this test bank are grouped together by chapter. Occasionally in multiple choice questions, a question can give students hints about the correct answer of another question. This is called cluing, and it can interfere with gaining an accurate view of the student’s knowledge. Efforts have been made to eliminate cluing within each chapter. However, cluing may be present for questions in different chapters. When selecting questions for a test, I recommend that instructors ensure that there is no cluing for items across chapters.
I wrote test questions with two assumptions: (1) students will not have any formulas memorized, and (2) students will not have access to any statistical tables. Instructors who require their students to memorize formulas or who will provide the statistical tables to students may find it advisable to adapt multiple choice questions into short answer questions. Some instructors may also wish to write hand calculation problems if that is an important learning outcome for their students. Additionally, some questions can be adapted to test students’ knowledge of similar concepts in other chapters.
Every chapter has at least 20 multiple choice questions and 5 short answer questions in the test bank. Short answer questions are always the last five listed for that chapter. Instructors who desire more short answer questions can adapt multiple choice questions into short answer questions. Instructors who desire calculation questions can adapt the Guided Practice examples in the textbook or end-of-chapter questions for this purpose. If an instructor does this, please ensure that students are provided with the relevant tables. Instructors who require students to memorize formulas should communicate this information clearly to students; other instructors should provide the relevant formulas with their test materials.
Correct answers are marked in bold.
Chapter 1
- The family of research methods that uses statistics to analyze numerical data is called
- Qualitative methods
- Statistical analysis
- Quantitative methods
- Descriptive statistics
- Which of the following is not a reason why students in the social sciences need to learn statistics?
- Evaluate the conclusions of researchers
- To help the social sciences obtain the same respect as the physical sciences
- Communicate findings to others
- Interpret research to create practical, real-world results
- Why do students and practitioners in the social sciences need to know how to separate good research from bad research?
- To know which articles should receive media attention and awards.
- To properly select articles to discuss with professors or teachers.
- The quality of published research varies greatly.
- All of the above.
- How do research questions and research hypotheses differ?
- Research questions are not scientific, while research hypotheses are scientific.
- Research hypotheses are appropriate for qualitative methods, while research questions are appropriate for qualitative methods.
- Research questions are part of confirmatory research, while research hypotheses are speculations about the world.
- Research hypotheses are expected beliefs the researcher has; research questions tend to be more exploratory.
- To be scientific, a research hypothesis must be
- Sophisticated
- Falsifiable
- Plausible
- Interesting
- In statistics, what is a population?
- Every person, event, or object that a researcher could wish to study.
- The particular people who are part of a study.
- All of the people who live in a country where the researchers conducts their study.
- None of the above.
- Why do researchers rarely have population data?
- It is unethical to force people to participate in a research study.
- Constraints of money and time may make it unfeasible to gather data from every population member.
- Many populations are too large.
- All of the above.
- A characteristic that is the same for all sample or populations members is a
- Variable
- Invariant characteristic
- Constant
- Consistent variable
- A dependent variable is the ________________ in the study.
- Constant
- Hypothesized cause
- Operationalization
- Outcome variable
- An independent variable is the ________________ in the study.
- Constant
- Hypothesized cause
- Operationalization
- Outcome variable
- In a _________________ study, a researcher does not manipulate or control the independent variable.
- Experimental
- Correlational
- Analytical
- Laboratory
- Which type of study has more applicability to real-life situations?
- Experimental
- Correlational
- Analytical
- Laboratory
- In a _________________ study, a researcher manipulates or controls the independent variable.
- Experimental
- Correlational
- Analytical
- Laboratory
- What is a common criticism of experimental studies?
- The studies are difficult to understand.
- They are too artificial.
- The data from experimental research is often not easy to interpret.
- All of the above.
- The purpose of __________________ statistics is to describe the data at hand.
- Descriptive
- Inferential
- Analytical
- Extrapolated
- What is a model?
- A simplification of a complex reality
- The summary of a study’s findings
- An understandable explanation of a study’s design
- All of the above
- The purpose of __________________ statistics is to estimate information about the population on the basis of sample data.
- Descriptive
- Inferential
- Analytical
- Extrapolated
- Why do social scientists often create models?
- To simulate a scenario and test whether a scientist’s theory is supported by data.
- Because researchers often wish to create a study that can serve as an example of future studies on the same topic.
- Reality is sometimes too messy and complicated to comprehend.
- All of the above.
- Which of the following is a type of model?
- Statistical models
- Theoretical models
- Visual models
- All of the above
- Which type of model is described in numerical terms or as an equation?
- Statistical models
- Theoretical models
- Visual models
- All of the above
- A bar graph, line graph, and chart are examples of
- Statistical models
- Theoretical models
- Visual models
- All of the above
- Theoretical models take the form of
- Verbal descriptions
- Numerical equations
- Graphs or charts
- None of the above
- A theory is
- A scientific law which has been shown to display validity across many studies that have been conducted throughout the world.
- A logically consistent system of principles that posit causal explanations for a wide range of phenomena in many contexts.
- A foundational idea that permits the design, execution, and interpretation of research within a scientific field.
- All of the above.
- Tony conducted a study where he collected data on two variables: the number of hours a person works and their job satisfaction. He believes that people with more satisfying jobs will choose to work more hours. This is an example of a
- Experimental study
- Analytical study
- Correlational study
- Laboratory study
- Tony conducted a study where he collected data on two variables: the number of hours a person works and their job satisfaction. He believes that people with more satisfying jobs will choose to work more hours. Which variable is the dependent variable?
- Job satisfaction
- Number of hours worked
- There is not enough information to choose a dependent variable.
- Both are dependent variables.
- If every model is wrong to some degree, explain how researchers judge a model.
- Models are judged by whether they are useful.
- Write a research hypothesis.
- Answers will vary, but the response should be a testable or falsifiable belief about a study’s outcome. It should not be in the form of a question.
- Why do researchers never say that their hypothesis or theory has been proven to be true?
- Because this would require a scientist to make every possible observation of a phenomenon. Otherwise, there could be a scenario or situation where the theory is disproven.
- If a population consists of college students in the social sciences, what would a constant among these individuals be?
- Answers will very, but student responses should be a characteristic of individuals that is the same for all population members (e.g., status as a college student, species that people belong to).
- Explain why demographic variables (e.g., race, sex) are considered independent variables.
- These variables cannot be outcomes—and therefore cannot be dependent variables. It makes much more theoretical sense that these could be theorized causes.
Chapter 2
- According to Stevens (1946, p. 677) measurement is:
- The application of scientific methods to obtain data
- The use of scientific instruments to measure objects
- The scientific assessment of objects
- The assignment of numbers to objects
- Data at the highest level of measurement:
- Has all qualities shared by the lower levels of measurement.
- Possesses an absolute zero point.
- Can be doubled without causing problems.
- All of the above.
- The process of defining a variable in a way that allows a researcher to collect numerical data about is called
- Operationalization
- Objectivism
- Reductionism
- Quantitative defining
- Data at Stevens’s highest level of measurement:
- Has all qualities shared by the lower levels of measurement.
- Possesses an absolute zero point.
- Can be doubled without causing problems.
- All of the above.
- Which option lists the four levels of data in ascending order (i.e., from lowest to highest)?
- Nominal, ordinal, ratio, interval
- Ratio, interval, ordinal, nominal
- Nominal, ordinal, interval, ratio
- Ratio, ordinal, interval, nominal
- According to Stevens (1946, p. 677) measurement is:
- The application of scientific methods to obtain data
- The use of carefully designed scientific instruments to measure objects
- The scientific assessment of objects using formal methods
- The assignment of numbers to objects or events according to rules
- Nominal data must be
- Able to preserve the ranking of subjects
- Mutually exclusive and exhaustive
- Different for different group members within the same group
- All of the above
- In nominal data, the numbers assigned to the categories
- Must be assigned in order so that larger groups receive larger numbers
- Are arbitrary
- Can never be negative
- Must reflect the rank order of the subjects
- Which of the following is not an acceptable mathematical function for nominal data?
- Counting
- Classification
- Calculating proportions
- Calculating averages
- In addition to the characteristics of nominal data, ordinal data must also have
- Rank order in the numbers
- Proportional representation of group members in the population
- Absolute zero
- Proportions calculated from scores
- Why is it always better to collect data at the highest level possible?
- Higher levels of data are easier to collect than lower levels of data.
- Lower levels of data require more preparation of the data before the statistical analysis can begin.
- Higher levels of data can always be converted down to lower levels.
- All of the above.
- Which mathematical procedures are acceptable for interval-level data, but not ordinal or ratio data?
- Calculating proportions
- Dividing to form averages
- Ranking subjects
- Dividing to form ratios
- What is the property that interval data have that ordinal data do not?
- Equal spacing between scale points
- Absolute zero
- Consistent data application rules
- Arbitrary numbers assigned to categories
- An absolute zero
- Is not present unless it is possible for a person in the sample to obtain a score of zero.
- Is required to calculate averages.
- Indicates the total absence of the quality being measured.
- All of the above.
- The number of movies that a person has seen in the past month is what type of data?
- Interval
- Nominal
- Ratio
- Ordinal
- Why is it that the Celsius temperature scale not a ratio-level scale?
- Negative numbers are possible (and meaningful) in the Celsius scale
- Because it lacks an absolute zero point
- The ratios that are formed with its numbers do not represent true ratios between temperatures of the amount of heat
- All of the above
- The level of data of ___________________________ is ambiguous.
- Rating scales
- Mental health variables
- Reaction time
- Group-level variables (e.g., a nation’s average education level)
- A(n) _________________ variable permits a wide range of scores that form a constant scale with no gaps at any point along the scale.
- Ratio
- Interval
- Dependent
- Continuous
- A(n) _________________ variable has a limited number of possible values and do not form a constant, uninterrupted scale of scores.
- Operationalized
- Discrete
- Ordinal
- Limited
- A sociologist asks her subjects their religious affiliation. What type of variable would this be?
- Continuous
- Interval
- Nominal
- Ordinal
- A researcher collects data on the length of individuals’ commute to their job. What type of variable is this?
- Ordinal
- Ratio
- Discrete
- Nominal
- Individuals learning a second language were labeled as “not proficient” (group 1), “basic proficiency” (group 2) “high proficiency” (group 3), and “fully proficient” (group 4). What type of data is this?
- Ordinal
- Nominal
- Continuous
- Ratio
- The Fahrenheit temperature scale is an example of a(n) ________________ variable.
- Ordinal
- Independent
- Interval
- Discrete
- What is the minimum level of data required to calculate proportions?
- Ratio
- Ordinal
- Nominal
- Interval
- What level of data can be used to rank order scores?
- Ratio
- Ordinal
- Interval
- All of the above
- Give an example of an operationalization.
- Answers will vary, but the response should be a method of defining a construct in a way that allows numerical data to be collected about it.
- What does it mean that categories in the data must be “mutually exclusive and exhaustive”?
- The categories are non-overlapping (mutually exclusive) and every sample member belongs to a category (exhaustive)
- Explain why test scores can be ordinal-, interval-, or ratio-level data, depending on the interpretation.
- If the score is the number of questions correctly, then the variable is ratio-level data. If the score is interpreted as the amount of a trait that the subject possesses, then the variable is interval-level data. If it is possible that there are not equal spaces or intervals between scores on a test, then the variable would be ordinal-level data.
- Explain what reductionism is and why it is a shortcoming of quantitative research.
- Reductionism is a philosophy that redefines variables so that they are a shallower version of the construct of interest. It is a problem of quantitative research because it means researchers don’t really study their constructs, but rather the operationalizations of the constructs.
- Why is it acceptable (and sometimes necessary) to create a “miscellaneous” or “other” category for data?
- Because categories must be exhaustive, which means that every sample member must belong to a category. Creating this “miscellaneous” category ensures that individuals who would otherwise not have a category can belong to one.
Chapter 3
- In a frequency table, all of the frequency values will add up to
- 0
- 10
- 100
- The sample size
- The width of a bar in a histogram is called the
- Response frequency
- Cumulative frequency
- Interval
- Ordinal width
- The height of a bar in a histogram corresponds to the
- Frequency of responses in that score range
- The cumulative frequency of responses in the score range
- The ordinal nature of histogram data
- None of the above
- What does a gap in the bars in a histogram represent?
- The discrete nature of the data in the histogram
- The presence of a score of zero, which indicates the variable is ratio-level data
- A platykurtic distribution of scores
- The absence of scores in the gap’s range
- Which of the following is a characteristic of a normal distribution?
- Unimodal
- Symmetrical
- Mean = Median = Mode
- All of the above
- The tails of a normal distribution
- Extend to from -∞ to +∞
- Contribute to the distribution’s platykurtic shape
- Are skewed
- All of the above
- Belgian mathematician Lambert Adolphe Jacques Quetelet was the first scientist to observe that a human characteristic was normally distributed. What variable did Quetelet measure?
- Intelligence test scores
- Height
- Reaction time
- Chest circumference
- What is the best method of determining whether a variable is normally distributed?
- Examine the data in a frequency table
- Create a histogram of the data
- Research how the variable is reported in previous studies
- All of the above
- Skewness measures the degree to which a distribution is
- Leptokurtic
- Made of interval-level data
- Continuous
- Non-symmetrical
- A histogram in which all the bars are the same height is called a
- Uniform distribution
- Positively skewed distribution
- Bimodal distribution
- Leptokurtic distribution
- An unskewed distribution has a skewness value that is
- Negative
- Zero
- Positive
- Undetermined
- Another term for a positively skewed distribution is
- Skewed left
- Skewed right
- Upwardly skewed
- Platykurtic
- The region of a distribution between the peak and the tail is called the
- Torso
- Midregion
- Midsection
- Shoulder
- What is the term for a distribution that has a taller peak, taller tails, and shorter shoulders than a normal distribution?
- Mesokurtic
- Platykurtic
- Leptokurtic
- Allokurtic
- A normal distribution is
- Mesokurtic
- Platykurtic
- Leptokurtic
- Allokurtic
- A platykurtic distribution generally
- Has a shape similar to a normal distribution
- Is more peaked and skinnier than a normal distribution
- Is flatter than a normal distribution
- Can vary in shape, according to the level of skewness in the distribution
- A leptokurtic distribution generally
- Has a shape similar to a normal distribution
- Is more peaked and skinnier than a normal distribution
- Is flatter than a normal distribution
- Can vary in shape, according to the level of skewness in the distribution
- In the histogram, if the peak is on the right side of the distribution, the data are:
- Positively skewed
- Negatively skewed
- Has skewness = 0.
- It is not possible to tell with this histogram.
- A leptokurtic distribution will have a kurtosis value that is
- Negative
- Zero
- Positive
- Undetermined
- Most distributions in the social sciences are
- Unimodal
- Bimodal
- Trimodal
- Platykurtic
- Visually, a histogram is most like a
- Infographic
- Pie chart
- Box plot
- Frequency polygon
- The center line in a box plot represents the
- Minimum
- Maximum
- Median
- Interquartile range
- The rectangle in a box plot represents the
- Minimum
- Maximum
- Median
- Interquartile range
- The advantage of a box plot over a histogram is that it can
- Show whether a distribution is bimodal
- Display distributions for multiple variables at once
- Display the kurtosis of a variable
- All of the above
- A stem-and-leaf plot
- Displays every score in the dataset
- Does not show frequently appearing scores in the dataset
- Is currently a more popular visual model than it was in the past
- All of the above
- Which type of visual model is best for showing trends over time?
- Frequency polygons
- Histograms
- Line graphs
- Bar graphs
- In a scatterplot, each dot represents a
- Variable
- Sample member
- Sample statistic
- Population parameter
- Which kind of data is appropriate for a frequency table?
- Ordinal
- Nominal
- Ratio
- All of the above
- If a distribution has a skewness value of -1.8 and a kurtosis value of -1.4, what would be an appropriate label for the distribution?
- Negatively skewed and mesokurtic
- Negatively skewed and platykurtic
- Unskewed and platykurtic
- Positively skewed and platykurtic
- If a distribution has a skewness value of +0.2 and a kurtosis value of +4.3, what would be an appropriate label for the distribution?
- Unskewed and leptokurtic
- Mesokurtic and positively skewed
- Platykurtic and unskewed
- Leptokurtic and negatively skewed
- If a distribution has a skewness value of -1.4 and a kurtosis value of +0.3, what would be an appropriate label for the distribution?
- Platykurtic and unskewed
- Unskewed and leptokurtic
- Unskewed and mesokurtic
- Negatively skewed and mesokurtic
- If a distribution has a skewness value of +2.3 and a kurtosis value of -2.1, what would be an appropriate label for the distribution?
- Positively skewed and mesokurtic
- Negatively skewed and leptokurtic
- Positively skewed and platykurtic
- Leptokurtic and negatively skewed
- If a distribution has a skewness value of -9.6 and a kurtosis value of +4.1, what would be an appropriate label for the distribution?
- Negatively skewed and leptokurtic
- Positively skewed and leptokurtic
- Negatively skewed and mesokurtic
- Positively skewed and mesokurtic
- What would be an appropriate label for the normal distribution?
- Mesokurtic, unskewed, and unimodal
- Leptokurtic, negatively skewed, and mesokurtic
- Platykurtic, unskewed, and unimodal
- Bimodal, positively skewed, and mesokurtic
- Draw a distribution that is platykurtic, negatively skewed, and unimodal.
- The distribution must be flat in shape, has a longer tail on the left, and only a single peak.
- Draw a distribution that is mesokurtic, positively skewed, and bimodal.
- The distribution must be as tall as a normal distribution, have a longer tail on the right, and two peaks.
- What is the cumulative frequency in a frequency table?
- Cumulative frequency is the number of sample members in a particular row in the frequency table, plus the number of individuals in all higher rows.
- A perfectly mesokurtic distribution will have a kurtosis value of:
- Zero
- Explain the differences between a bar graph and a histogram.
- A bar graph is for nominal/discrete data and always has gaps between bars, while a histogram is designed for continuous (or interval-data or ratio-) data and gaps show score intervals where there are no scores in the data.
Chapter 4
- Models of central tendency are used to
- Move extreme scores more towards the center of the distribution
- Describe where the middle of the histogram lies
- Measure the degree to which the center of the sample reflects the center of the population
- Display data in an understandable manner
- Which measure of central tendency can be calculated for all levels of data?
- Mean
- Median
- Mode
- Trimmed mean
- Which measure of central tendency is most influenced by extreme values?
- Mean
- Median
- Mode
- Trimmed mean
- A drawback of the trimmed mean is that
- It requires eliminating scores from the dataset.
- The trimmed mean requires ordinal data to calculate.
- It does not guarantee the removal of both high and low outliers.
- All of the above.
- The interquartile range is a model of variability that is most like which model of central tendency?
- Mode
- Median
- Mean
- Trimmed mean
- Calculating the 10% trimmed mean requires using how much of data?
- 10%
- 80%
- 90%
- 110%
- In statistics, n is the abbreviation for
- Numbers
- Sample size
- Median
- Trimmed mean
- An outlier is a score that is unusual because it is
- A score that needs to be removed from a dataset before analysis can occur
- Much higher or much lower than the other scores in the dataset
- Not a score that belongs to population members
- A duplicate score that distorts the data
- What is an advantage of the mode?
- It can be used in later statistical calculations
- It is not distorted by outliers
- It corresponds to the center number in the distribution
- All of the above
- Which of the following samples would consist entirely of outliers?
- Children who watch cartoons
- People who catch influenza
- Nobel Prize winners
- None of the above
- The median is defined as the _______________ of a distribution.
- Average score
- Standard deviation
- Most frequent score
- Middle score
- The advantage of the sample mean is that it is
- Easily understandable
- The best estimate of the population mean
- The average score in the sample
- All of the above
- The mean is defined as the score that
- Exceeds half the scores in the dataset
- The deviation scores of the sample sum to zero
- Is the most common score in the dataset
- None of the above
- One disadvantage of the mean is that it
- Maximizes the deviation scores
- Cannot be used in later statistical calculations
- May not be a score in the dataset
- All of the above
- What is the impact that outliers have on the mean?
- They pull the mean closer to their scores
- Outliers have no impact on the mean
- They move the mean closer to the middle of the distribution
- Outliers move the mean closer to the mode by making the distribution more skewed
- A deviation score is defined as
- The sum of all scores in a dataset, minus the median
- The amount of sampling error in a score
- The most common score in the dataset
- A score minus the mean
- What is a disadvantage of the trimmed mean?
- It does not produce a realistic estimate of central tendency
- It cannot eliminate the influence of outliers
- It may not be appropriate to eliminate data from the dataset
- All of the above
- When models of central tendency differ, what is the appropriate course of action?
- Find the average value of the different models to produce a more accurate estimate of central tendency
- Use the mean as a model of central tendency because it is the most accurate model of central tendency
- Re-calculate the central tendency values to determine which one was in error
- Choose a preferred model and justify the decision
- What is a statistical assumption?
- A requirement that the data must meet in order for a statistical procedure to be used
- An underlying belief about the nature and purpose of a statistical procedure
- The assumed population parameter that the sample statistic is supposed to estimate
- A formula used to calculate a statistic
- What is the procedure for calculating the range of data?
- Divide the mean by the highest score in the dataset
- Subtract the lowest score from the highest score
- Subtract 1 from the highest score
- None of the above
- The interquartile range is defined as
- The range of the middle 50% of scores
- The third quartile minus the first quartile
- The 75th percentile minus the 25th percentile
- All of the above
- Another term for the median is the
- Derived central tendency
- Second quartile
- Most central score
- Distributional midpoint
- Which model of variability will be produce the largest value?
- Range
- Interquartile range
- Standard deviation
- It depends on the dataset
- Which formula is appropriate for calculating a standard deviation when a researcher has sample data and does not wish to generalize to the population?
- Any of the above equations is appropriate
- In the equation , the n – 1 is called a(n)
- Compensatory correction
- Underestimate correction
- Bessel’s correction
- Sampling adjustment
- In the equation , why is it necessary to put n – 1 in the denominator?
- Because having n in the denominator would be appropriate for population data
- To convert the standard deviation to the variance
- Using n in the denominator would result in a population standard deviation estimate that is too small
- All of the above
- Which level of measurement is most basic level required to calculate the mean?
- Nominal
- Ordinal
- Interval
- Ratio
- Which level of measurement is most basic level required to calculate the median?
- Nominal
- Ordinal
- Interval
- Ratio
- Which level of measurement is most basic level required to calculate the mode?
- Nominal
- Ordinal
- Interval
- Ratio
- Which level of measurement is most basic level required to calculate the range?
- Nominal
- Ordinal
- Interval
- Ratio
- Which level of measurement is most basic level required to calculate the interquartile range?
- Nominal
- Ordinal
- Interval
- Ratio
- Which level of measurement is most basic level required to calculate the trimmed mean?
- Nominal
- Ordinal
- Interval
- Ratio
- Which level of measurement is most basic level required to calculate the standard deviation?
- Nominal
- Ordinal
- Interval
- Ratio
- Which level of measurement is most basic level required to calculate the variance?
- Nominal
- Ordinal
- Interval
- Ratio
- What is the relationship between models of central tendency and models of variability?
- Models of variability verify the results from models of central tendency
- They provide complementary information needed to understand the distribution
- Models of central tendency provide limits to the possible values of models of variability
- All of the above
- If the variance of a dataset is equal to zero, this indicates that
- All the scores in the data are equal to one another
- The standard deviation is also equal to zero
- There is no variability in the data
- All of the above
- Explain how to calculate a 5% trimmed mean.
- Remove the highest 5% and lowest 5% of scores and calculate the mean of the remaining of scores (i.e., the middle 90% of scores).
- Name an advantage of keeping outliers in a sample.
- Studying outliers can provide information that can help more typical individuals
- Also acceptable: samples that consist of outliers may be interesting.
- How do you calculate a median where there is an even number of scores?
- Find the middle two scores, add them together, and divide by two.
- Also acceptable: find the number halfway between the middle two scores.
- What is the relationship between the standard deviation and the variance?
- The standard deviation is the square root of the variance.
- Also acceptable: the variance is equal to the standard deviation squared.
- Why is it important to calculate models of both central tendency and variability?
- Both types of models provide different, but complementary, information.
Chapter 5
- What is the purpose of a linear transformation?
- To remove outliers from a sample
- To display the data and compare it to a normal distribution
- To convert a set of scores to a new scale
- All of the above
- Adding a constant to every score will have what impact on the mean?
- It will decrease the mean by the same amount
- It will increase the mean by the same amount
- It will not change the mean
- The impact of this linear transformation on the mean varies
- Subtracting a constant from every score will have what impact on the mean?
- It will decrease the mean by the same amount
- It will increase the mean by the same amount
- It will not change the mean
- The impact of this linear transformation on the mean varies
- Adding a constant to every score will have what impact on the median?
- It will decrease the median by the same amount
- It will increase the median by the same amount
- It will not change the median
- The impact of this linear transformation on the median varies
- Subtracting a constant from every score will have what impact on the median?
- It will decrease the median by the same amount
- It will increase the median by the same amount
- It will not change the median
- The impact of this linear transformation on the median varies
- Adding a constant to every score will have what impact on the trimmed mean?
- It will decrease the trimmed mean by the same amount
- It will increase the trimmed mean by the same amount
- It will not change the trimmed mean
- The impact of this linear transformation on the trimmed mean varies
- Subtracting a constant from every score will have what impact on the trimmed mean?
- It will decrease the trimmed mean by the same amount
- It will increase the trimmed mean by the same amount
- It will not change the trimmed mean
- The impact of this linear transformation on the trimmed mean varies
- Adding a constant to every score will have what impact on the range?
- It will decrease the range by the same amount
- It will increase the range by the same amount
- It will not change the range
- The impact of this linear transformation on the range varies
- Subtracting a constant from every score will have what impact on the range?
- It will decrease the range by the same amount
- It will increase the range by the same amount
- It will not change the range
- The impact of this linear transformation on the range varies
- Adding a constant to every score will have what impact on the interquartile range?
- It will decrease the interquartile range by the same amount
- It will increase the interquartile range by the same amount
- It will not change the interquartile range
- The impact of this linear transformation on the interquartile range varies
- Subtracting a constant from every score will have what impact on the interquartile range?
- It will decrease the interquartile range by the same amount
- It will increase the interquartile range by the same amount
- It will not change the interquartile range
- The impact of this linear transformation on the interquartile range varies
- Adding a constant to every score will have what impact on the standard deviation?
- It will decrease the standard deviation by the same amount
- It will increase the standard deviation by the same amount
- It will not change the standard deviation range
- The impact of this linear transformation on the standard deviation varies
- Subtracting a constant from every score will have what impact on the interquartile range?
- It will decrease the standard deviation by the same amount
- It will increase the standard deviation by the same amount
- It will not change the standard deviation
- The impact of this linear transformation on the standard deviation varies
- Adding a constant to every score will have what impact on the variance?
- It will decrease the variance by the same amount
- It will increase the variance by the same amount
- It will not change the variance
- The impact of this linear transformation on the variance varies
- Subtracting a constant from every score will have what impact on the variance?
- It will decrease the variance by the same amount
- It will increase the variance range by the same amount
- It will not change the variance
- The impact of this linear transformation on the variance varies
- Multiplying every score by a constant will change the ______________________ by multiplying the statistic by the same constant?
- Median
- Mean
- Trimmed mean
- All of the above
- Dividing every score by a constant will change the ______________________ by dividing the statistic by the same constant?
- Median
- Mean
- Trimmed mean
- All of the above
- Multiplying every score by a constant will change the ______________________ by multiplying the statistic by the same constant?
- Interquartile range
- Range
- Standard deviation
- All of the above
- Dividing every score by a constant will change the ______________________ by dividing the statistic by the same constant?
- Interquartile range
- Range
- Standard deviation
- All of the above
- If a sample member has an original score less than the mean, you can predict that the individual’s z-score will be
- Negative
- Zero
- Positive
- It is impossible to tell
- A linear transformation changes the _______________ of a histogram
- Shape
- Skewness
- Axis
- Kurtosis
- A set of z-scores has a mean of _________ and a standard deviation of _________.
- 0, 1
- 100, 10
- 10, 1
- 1, 0.1
- Which z-score is closest to the mean?
- -3.2
- 1.5
- 1.9
- -1.8
- Which z-score is furthest from the mean?
- -3.2
- 1.5
- 1.9
- -1.8
- Which of the following is a benefit of z-scores?
- A set of z-scores will always be normally distributed
- They permit comparisons of scores across different scales
- Calculating z-scores will convert ordinal data to interval data
- All of the above
- Another term for z-scores is
- Normalized scores
- Standardized scores
- Distributional scores
- Zero-centered scores
- In a normal distribution, half of z-scores are below what value?
- -1
- 0
- .50
- 10
- In a normal distribution, 2.28% of scores will be above what value?
- -15
- -2
- 1
- 2
- If the mean of a set of scores is 13 and the standard deviation is 4, what is the original score for someone who had a z-score of +1.5?
- 7
- 9
- 17
- 19
- If the mean of a set of scores is 9 and the standard deviation is 6, what is the original score for someone who had a z-score of -2.5?
- -6
- 6
- 15
- 24
- When data are converted to z-scores, what happens to the shape of the distribution?
- There are no changes in the shape of the histogram.
- In a normal distribution, almost all (99.73%) of z-scores will be between which two values?
- -3 and 3
- Interpret a z-score of +1.6.
- The person’s original score is 1.6 standard deviations above the mean of the dataset.
- If every score in a dataset if multiplied by 2, explain how you can estimate the variance of the scores after this linear transformation.
- The variance can be calculated by multiplying the standard deviation by 2 and squaring the result.
- Explain why the z-score formula is a type of linear transformation.
- The z-score formula subtracts the same number (in this case, the mean) from every score and then divides the result by the same number (which is the standard deviation). Both of these numbers are constants.
Chapter 6
- Probability is the branch of mathematics that is concerned with the
- Result of a specific event that that occurs in nature
- Likelihood of outcomes for an event
- The combinations of independent and dependent variable values that occur in a study
- All of the above
- _______________ probability is calculated by observing trials and their outcomes.
- Analytical
- Theoretical
- Empirical
- Calculated
- _______________ probability is estimated on the basis of theory.
- Analytical
- Theoretical
- Empirical
- Calculated
- A probability of 1.00 indicates that
- The outcome is certain to happen
- The outcome is equally likely as any other outcome
- The outcome’s likelihood is uncertain
- The probability’s interpretation is uncertain
- _______________ is calculated in the same way as relative frequency.
- Theoretical probability
- Joint probability
- Empirical probability
- None of the above
- Independent trials are
- Scores collected for an independent variable
- Separate results for the different probability calculations
- A causal impact on probability results
- Trials that do not impact the outcome of any other trial
- When calculating the likelihood that either of two possible outcomes could occur in a single trial, one should
- Add together the probabilities for each outcome
- Subtract the lesser probability from the greater probability
- Multiply the probabilities for each outcome together
- Use the probability of the more likely outcome
- When calculating the likelihood that two outcomes will occur simultaneously, one should
- Add together the probabilities for each outcome
- Subtract the lesser probability from the greater probability
- Multiply the probabilities for each outcome together
- Use the probability of the more likely outcome
- A probability distribution is
- Normal distribution that has been converted to a joint probability
- Histogram of probabilities for different outcomes
- A uniform distribution of probabilities
- The most likely histogram for a dataset.
- Which of the following is an example of the gambler’s fallacy?
- “My lucky rabbit’s foot will help me win this game of blackjack.”
- “I got a good parking spot at work every day this week, and that streak will continue.”
- “It has been a long time since I’ve caught a fish, so I’m due for one.”
- “I have just as many good days as bad days, so a good day is possible tomorrow.”
- Which of the following is an example of the hot hand fallacy?
- “My lucky rabbit’s foot will help me win this game of blackjack.”
- “I got a good parking spot at work every day this week, and that streak will continue.”
- “It has been a long time since I’ve caught a fish, so I’m due for one.”
- “I have just as many good days as bad days, so a good day is possible tomorrow.”
- In a probability distribution, what does the height of a bar represent?
- The type of probability that the researcher calculated
- The most likely outcome for an individual
- The probability of an outcome
- The shape of the histogram that a variable produced
- What happens to an empirical probability distribution as the number of trials increases?
- The empirical probability distribution better resembles the theoretical probability distribution
- The empirical probability becomes more spread out as the number of trials increases
- The empirical probability distribution becomes more uniform
- The empirical probability distribution becomes less variable as the number of trials increases
- When an outcome is determined by the sum of independent trials, what shape is the theoretical probability distribution?
- Uniform
- Negatively skewed
- Positively skewed
- Normal
- What is the process of drawing conclusions about populations based on sample data?
- Generalization
- Descriptive statistics
- Extrapolation
- Evidentiary procedure
- A histogram of sample statistics (e.g., means) is called a
- Statistic distribution
- Sampling distribution
- Standard error
- Sampling frequency histogram
- A distribution of an infinite number of sample means (n of each sample > 25) will be normally distributed if the original population distribution is
- Normal
- Uniform
- Leptokurtic
- Any shape
- With a large number of sample means, the best estimate of the population mean is
- The variance sample mean values
- The standard deviation of sample mean values
- The mean of sample mean values
- The largest sample mean value
- The standard deviation of a series of sample means is the
- Population standard deviation
- Standard error
- Sampling error
- Variance parameter
- When calculating means from a large number of samples, __________________ will produce a smaller sampling error.
- More consistent means
- Less consistent means
- Larger means
- Smaller means
- To reduce sampling error, a researcher should
- Sample from a normally distributed population
- Reduce the standard deviation of the sample means
- Select samples that have fewer outliers
- Have a larger sample size
- Populations with high variability will produce _____________ sampling error than populations with low variability.
- More
- Equal
- Less
- More inconsistent
- The ____________________ will always have a larger standard deviation than the distribution of means.
- Individual sample histograms
- Theoretical sampling distribution
- Population distribution
- Empirical sampling distribution
- Explain the Central Limit Theorem.
- The Central Limit Theorem states that a sampling distribution of means will be normally distributed if each sample will have an n of 25 or more.
- Explain why the population standard deviation will always be smaller than the standard deviation of sample means.
- Means are less variable than raw data because any outliers in a particular sample will usually have their influence lessened by the other scores in the sample. As a result, a collection of means will have less variability than raw data.
- Why does the Central Limit Theorem permit researchers to use a theoretical probability distribution when generalizing, instead of needing to use an empirical probability distribution?
- Because a distribution of an infinite number of sample means will be normally distributed (as long as n is 25 or more per sample) it is not necessary to create an empirical distribution because a researcher already knows the ultimate shape of the sampling distribution.
- Explain what sampling error is.
- Sampling error is the normal variation in statistics across samples.
- List an example of when a joint probability would be appropriate to calculate.
- Answers will vary, but student responses should refer to the probability that two specific outcomes will occur in succession or simultaneously.
Chapter 7
- The z-test is
- A test of equality of two sample means
- A test of whether the sample and population standard deviations are equal
- The only statistical test that does not require a null hypothesis
- The simplest null hypothesis statistical significance test
- The purpose of a z-test is to determine whether
- The null hypothesis fits the data
- The sample is typical of the population
- The sample mean and population means are equal
- All of the above
- What is the null hypothesis of a z-test?
- sx = σx
- = 0
- If the null hypothesis is rejected, this indicates that the null hypothesis
- Does not fit the data well
- Was incorrect
- Requires an additional sample for verification
- All of the above
- The α value indicates the
- Location of the rejection region
- Decision to reject the null hypothesis is correct
- Size of the rejection region
- Appropriate same size for a null hypothesis statistical significance test
- The shape of a sampling distribution in a z-test is
- Uniform
- Normal
- Platykurtic
- Different, depending on the sample size
- The groups in a z-test are the
- Sample and population
- Means and standard deviations
- Higher scoring individuals and lower scoring individuals
- Outliers and the other sample members
- Why do most researchers test the null hypothesis if they do not really believe it?
- The null hypothesis is the most logical hypothesis to test
- Disproving the null hypothesis can provide support for a researcher’s theory
- A researcher cannot prove their hypothesis to be true, but they can disprove a null hypothesis
- All of the above
- What are the limits to the values of α?
- -1 to 1
- 0 to 1
- 1 to 10
- 1 to 100
- A one-tailed z-test is appropriate when
- The data indicate that it would be easier to reject the null hypothesis with a one-tailed test than with a two-tailed test
- The rejection region is smaller than usual
- The researcher has no prior expectations of the results of the study
- Prior research leads an expectation of the results of the study
- A two-tailed z-test is appropriate when
- The data indicate that it would be easier to reject the null hypothesis with a one-tailed test than with a two-tailed test
- The rejection region is smaller than usual
- The researcher has no prior expectations of the results of the study
- Prior research leads an expectation of the results of the study
- What is the alternative hypothesis for a two-tailed z-test?
- The purpose of an alternative hypothesis is to
- Provide a hypothesis for the researcher to believe if the null hypothesis is rejected
- Give the null hypothesis a different form of expression
- Find results that can support a research hypothesis
- All of the above
- Which alternative hypothesis is appropriate for a one-tailed z-test where the rejection region is on the left side of the sampling distribution?
- Which alternative hypothesis is appropriate for a one-tailed z-test where the rejection region is on the right side of the sampling distribution?
- The critical value is defined as the location on the sampling distribution where
- The sample mean’s location in relationship to the population mean
- The rejection region begins
- The distribution becomes normal
- None of the above
- The z-observed value is obtained by
- Using a formula to calculate it
- Dividing the α in half
- Calculating the effect size
- Looking it up in a table
- In the seventh step of a z-test, it is appropriate to reject the null hypothesis if
- If the z-critical value is closer to the middle of the distribution than the z-critical value
- The α value is less than the p-value
- The z-observed value is beyond the z-critical value
- The rejection region is larger than the sampling distribution
- The z-observed formula most closely resembles which other formula?
- The standard deviation formula
- The standard error formula
- The mean formula
- The z-score formula
- The effect size for a z-test is
- A measure of how poorly the null hypothesis fits the data
- Related to the z-score formula
- A measure of the difference between sample and population means
- All of the above
- If the null hypothesis is a perfect model for the data, the effect size for a z-test will be
- 100%
- 0
- -1
- 10
- In theory, the limits of Cohen’s d are
- 0 and 1
- -∞ to +∞
- -3 and +3
- 0 and 100
- A Type I error occurs when
- A researcher rejects the null hypothesis when the null hypothesis is actually true
- A researcher makes a calculation error when conducting a null hypothesis statistical significance test
- The null hypothesis is incorrectly defined by the researcher
- A researcher retains the null hypothesis when the null hypothesis is actually false
- A Type II error occurs when
- A researcher rejects the null hypothesis when the null hypothesis is actually true
- A researcher makes a calculation error when conducting a null hypothesis statistical significance test
- The null hypothesis is incorrectly defined by the researcher
- A researcher retains the null hypothesis when the null hypothesis is actually false
- Which of the following is true of Type I and Type II errors?
- It is impossible to commit both errors at the same time
- It is never possible to know whether one has committed an error
- The p-value is the probability of committing a Type I error if the null hypothesis were perfectly true
- All of the above
- On a sampling distribution, a p-value can be represented as
- The area between the critical value and the middle of the distribution
- The area between the observed value and the middle of the distribution
- The area beyond the observed value
- The area beyond the critical value
- A p-value can never equal zero because
- The p-value is calculated with only negative numbers
- It would indicate that the sample could have never come from the population
- A zero would indicate that the null hypothesis is a perfect model for the data—which never happens in real life
- The limits of p-values are from 1 to 100.
- The value of ______________ is always arbitrary.
- Cohen’s d
- The z-observed value
- α
- The null hypothesis
- As sample sizes get larger, it becomes ________________ to reject the null hypothesis
- Easier
- More difficult
- Less sensible
- More intuitive
- In statistics, a special case is
- A specific application of the General Linear Model to data
- The example that a statistical procedure’s creator used to demonstrate it
- A situation where the assumptions of a statistical method break down and do not apply
- A simplification of a more complex statistical method
- What is the advantage to learning the General Linear Model as a statistics student?
- The GLM provides a framework for understanding statistics
- Each GLM procedure is just a modified version of previous procedures
- It makes new statistical procedures easier to understand
- All of the above
- What are the three characteristics of General Linear Model (GLM) procedures?
- (1) All GLM procedures apply weights to data during the analysis process. (2) All GLM procedures examine the relationships between independent and dependent variables. (3) All GLM procedures produce an effect size.
- Why is a rejection region in the tails of a z-distribution?
- Because this is where the least likely sample means would be if a null hypothesis were true. Therefore, we put the rejection region in the tails because sample means this range fit the null hypothesis very poorly and indicate that it should be rejected.
- Provide an example where a one-tailed z-test would be appropriate and explain why the test is appropriate.
- Responses will vary, but the student’s example indicates that previous theory or research leads them to expect a result of the study.
- A z-table sometimes does not display a value that you need. What should you do if the table does not contain the exact value indicated by your data?
- Find the value highest value in your table that does not exceed the exact desired value in the dataset.
- Also acceptable: Use the Price Is Right rule.
- What information does Cohen’s d provide that a z-test does not?
- Cohen’s d quantifies the size of the difference between means, whereas a z-test merely describes whether the null hypothesis fits the data or not.
Chapter 8
- When conducting a one-sample t-test, why is it necessary to make some adjustments to the z-test procedures?
- Because the one-sample t-test is not part of the General Linear Model
- The population distribution shape is not known
- Because σ is not known
- All of the above
- The groups in a one-sample t-test are the
- Sample and population
- Means and standard deviations
- Higher scoring individuals and lower scoring individuals
- Outliers and the other sample members
- What is the null hypothesis of a one-sample t-test?
- sx = σx
- = 0
- The default α value in a one-sample t-test is
- 0
- .01
- .05
- .50
- What is the alternative hypothesis for a two-tailed one-sample t-test?
- Which alternative hypothesis is appropriate for a one-tailed one-sample t-test where the rejection region is on the left side of the sampling distribution?
- Which alternative hypothesis is appropriate for a one-tailed one-sample t-test where the rejection region is on the right side of the sampling distribution?
- HARKing stands for
- Hypothesizing after results are known
- Histogram with high and random kurtosis
- Hypotheses that alter recent knowledge
- Histogram and ratio data kurtosis
- Which of the following is an example of HARKing?
- A researcher using the statistical results to alter their beliefs about the phenomenon
- Searching for an explanation of a study’s results
- Changing a two-tailed test into a one-tailed test in order for to make it easier to reject the null hypothesis
- Completing a study and then using those results to plan a new study
- What are degrees of freedom?
- The differences between the sample standard deviation value and the population standard deviation value
- The number of data points that can vary, given the statistical results for the dataset
- The difference between the t-observed and the t-critical values
- The difference between the results of a z-test and the results of a one-sample test for the same data
- In the seventh step of a one-sample t-test, it is appropriate to retain the null hypothesis if
- If the z-critical value is closer to the middle of the distribution than the z-critical value
- The α value is greater than the p-value
- The t-observed value is beyond the t-critical value
- The rejection region is larger than the sampling distribution
- A Cohen’s d value of ______________ is approximately the smallest effect size that would be noticeable to an attentive observer in daily life.
- 0.00
- 0.20
- 0.50
- 1.00
- In a one-sample t-test, Cohen’s d quantifies the
- Number of degrees of freedom in the sample
- Importance of a finding
- Difference between two group means
- Statistical significance of the one-sample t-test
- Comparing effect sizes allows researchers to
- Ascertain which variables have the largest group differences
- Determine the most effective treatment for a condition
- Help understand the relationship between group differences on different variables
- All of the above
- Under what condition does a one-sample t-test simplify to a z-test?
- When a variable is normally distributed
- When the degrees of freedom is equal to infinity
- When the sample is representative of the population
- All of the above
- When the p-value for a one-sample t-test is less than the α value,
- The t-observed value is closer to zero than the t-critical value is
- The null hypothesis should be rejected
- A Type II error has occurred
- The null hypothesis should be retained
- If using t-table in Appendix A2, each of the following is a possible p-value range, except
- p < .05
- p > .05
- .01 > p > .05
- p < .01
- If the results of a one-sample t-test are statistically significant, this indicates that
- The statistical results are important
- The null hypothesis was rejected
- The results can be translated into a treatment to help people
- There is no difference between the sample and population means
- What is practical significance?
- A result that are almost statistically significant
- A result that indicates that fewer sample members need treatment than before
- A group difference that is practically useful in the real world
- All of the above
- What is clinical significance?
- A result that are almost statistically significant
- A result that indicates that fewer sample members need treatment than before
- A group difference that is practically useful in the real world
- All of the above
- Confidence intervals describe
- A range of plausible values for a population parameter
- The degree of confidence that a researcher has in their quality of their sample
- The values that cannot be population parameters
- The degree to which future samples will replicate the findings
- A point estimate of a population parameter is
- Inside the confidence interval
- The best estimate that a sample produces for a population parameter
- Usually not precisely correct
- All of the above
- The minimum value of the confidence interval is called the
- Bottom
- CI minimum
- Lower limit
- Lower threshold
- The maximum value of the confidence interval is called the
- Top
- CI maximum
- Upper limit
- Higher threshold
- A narrow confidence interval indicates
- That there is too much sampling error in the population
- That the point estimate is likely trustworthy
- The one-sample t-test cannot produce interpretable results
- The sample size is small
- A 95% confidence interval indicates that
- There is a 95% chance that a study’s results would replicate in later samples
- We would be 95% sure that the population parameter is inside the confidence interval
- 95% of later sample means would fall within the confidence interval
- 95% of confidence intervals from repeated samples would include the population parameter
- A confidence interval indicates the level of _________________ in the sample.
- Sampling error
- Effect size uncertainty
- Mean error
- Capture percentage
- Confidence intervals can also be used to
- Verify the results of the Central Limit Theorem
- Conduct a null hypothesis statistical significance test
- Determine the shape of the population distribution
- Avoid the possibility of a Type I error
- In addition to the population mean, the one-sample t-test can examine whether there is a difference between the sample mean and
- The sample standard deviation
- The population standard deviation
- Any number the researcher chooses
- All of the above
- If the null hypothesis in a one-sample t-test is rejected when α = .01, what will happen when α is changed to .05?
- The null hypothesis will be retained
- The null hypothesis will be rejected
- The one-sample t-test will be converted into a z-test
- It is impossible to tell
- Describe how to use a t-table to find a range of p-values for a one-sample t-test.
- A p-value range can be found by examining which alpha values the null hypothesis is rejected or retained for. If the null hypothesis has been rejected, then p is lower than alpha. If the null hypothesis has been retained, then p is higher than alpha.
- What does the term “statistical significance” indicate?
- “Statistical significance” indicates that the null hypothesis has been rejected.
- Give an example of how the concepts of practical significance or clinical significance can sometimes be more useful than statistical significance.
- Responses will vary, but student examples should be situations where the context of importance or the clinical usefulness (e.g., whether people actually are well enough to not need treatment) is important.
- Explain why a point estimate is always inside the confidence interval.
- Because the confidence interval is the range of plausible values for a population parameter, while the point estimate is the best estimate of the parameter. The best estimate will always be plausible—and therefore inside the confidence interval.
- Explain why a researcher might use a one-sample t-test to compare a sample mean to a number that is not the population mean.
- Sometimes a researcher might want to compare a sample mean to another number because that other number might have an important interpretation (e.g., for diagnosis) and it would be important to determine whether the sample mean is equal to—or not equal to—that number.
Chapter 9
- Which of the following is an example of paired data?
- Scores from child participants and scores from adult participants
- Two scores collected from the same individual
- Scores from native English speakers and scores from native Spanish speakers
- All of the above
- Which of the following is an example of paired data?
- A set of scores from older adults and scores from younger adults
- Scores from a population of religious people and scores from a population of non-religious people
- Scores from a sample of parents and scores from each’s parent’s child
- All of the above
- What is the example of an artificial pairing of scores?
- Scores connected to people’s therapeutic treatments
- A researcher pairing up two similar people in different groups
- Two scores collected by a computer from the same person
- All of the above
- What is the null hypothesis for a paired-samples t-test?
- = 0
- > 0
- < 0
- ≠ 0
- What is the alternative hypothesis for a two-tailed paired-samples t-test?
- = 0
- > 0
- < 0
- ≠ 0
- Which alternative hypothesis would be appropriate for a one-tailed paired-samples t-test when all of the rejection region is on the right side of the distribution?
- = 0
- > 0
- < 0
- ≠ 0
- Which alternative hypothesis would be appropriate for a one-tailed paired-samples t-test when all of the rejection region is on the left side of the distribution?
- = 0
- > 0
- < 0
- ≠ 0
- In a paired-samples t-test the sampling distribution is made up of
- Differences between sample means
- Sample means
- Population means
- Sample standard deviations
- As the sample size increases in a paired-samples t-test increases, the sampling distribution more closely resembles a
- Uniform distribution
- Normal distribution
- Leptokurtic distribution
- Skewed distribution
- Glass’s Δ is appropriate when
- A treatment could change the standard deviation of scores
- The null hypothesis in a paired-samples t-test has been rejected
- The null hypothesis in a paired-samples t-test has been retained
- The effect size is greater than 1.0
- The pooled standard deviation is
- The standard deviation of the effect size
- Another name for the variance of a sample
- The overall standard deviation of multiple groups combined
- The population standard deviation estimate
- Because there are three effect size options for a paired-samples t-test, a researcher must
- Justify the choice of effect size
- Clearly explain which effect size option you choose
- Understand the differences between the effect size options
- All of the above
- Which is more serious, a Type I or a Type II error?
- Type I error
- Type II error
- Both errors are equally serious
- It depends on the situation
- Why do the one-sample t-test and the paired samples t-test share so many similarities?
- Both are members of the General Linear Model
- The sampling distributions for both procedures are the same shape
- A paired samples t-test uses differences scores to conduct a one-sample t-test
- All of the above
- Which two characteristics make it easier to reject the null hypothesis?
- A higher α and larger sample size
- A higher α and smaller sample size
- A lower α and larger sample size
- A lower α and a smaller sample size
- The n used to calculate the t-observed value in a paired-samples t-test refers to the
- Total number of scores
- Number of pairs of scores
- Number of degrees of freedom
- Total number of scores minus 1
- If a paired-samples t-test result is statistically insignificant, this means that
- The p-value is greater than the α value
- The null hypothesis was retained
- The t-observed value is closer to zero than the t-critical value is
- All of the above
- Why is it is never possible to know with complete certainty whether a Type I or Type II error has occurred?
- The sample sizes are usually too small to conclusively produce a result about the null hypothesis
- Because we never fully know whether the null hypothesis is true or not
- The paired-samples t-test is not a suitable method for investigating Type I or Type II error
- None of the above
- The formula for Glass’s Δ is most like the formula for
- Cohen’s d
- The sample mean
- The sample standard deviation
- The t-observed value
- In a paired-samples t-test, the number of scores in each group is
- Larger in the first group
- 25 or more in both groups
- Equal in the two groups
- Larger in the second group
- How are the two sets of scores in a paired samples t-test prepared for analysis?
- By subtracting one score in a pair from the other to produce a difference score
- Why does Glass’s Δ use the control group’s standard deviation to compare the means for the two groups in a t-test?
- Because often interventions can change the standard deviation of the data and make individuals more variable. This would reduce the effect size, so it may make more sense to compare groups using a standard deviation of the control group because they will usually better resemble the population. This is because the population will usually not have been exposed to the treatment.
- What is the impact of decreasing alpha from .05 to .01 on the probability of committing a Type II error?
- Decreasing alpha will always increase the probability of a Type II error because it makes it harder to reject a null hypothesis.
- Provide an example of a situation where a researcher can tolerate high Type II error but wants to minimize a Type I error as much as possible.
- Responses will vary, but student examples should be situations where a Type I error/false positive is a very costly/unpleasant/undesirable outcome, and a Type II error/false negative is comparatively less serious or damaging.
- In a z-test and a one-sample t-test, the sampling distribution consisted of sample means. However, this is not true for a paired samples t-test. What does the sampling distribution of a paired samples t-test consist of?
- The sampling distribution of a paired samples t-test consists of differences between means.
Chapter 10
- What is the null hypothesis in an unpaired two-sample t-test?
- What level of data is the independent variable in an unpaired two-sample t-test?
- Interval
- Ordinal
- Nominal
- Ratio
- Which of the following is the alternative hypothesis for a two-tailed test in an unpaired two-sample t-test?
- Which of the following is the alternative hypothesis for a one-tailed test in an unpaired two-sample t-test where the rejection region is on the left?
- Which of the following is the alternative hypothesis for a one-tailed test in an unpaired two-sample t-test where the rejection region is on the right?
- A _____________________ unpaired two-samples t-test is appropriate for situations where there is little prior research or theory guiding expectations of results.
- Unpaired two-sample t-test
- One-tailed test
- Two-tailed test
- None of the above
- What is the pooled variance?
- The square of the pooled standard deviation
- The variance of scores from two samples that are combined together
- Part of the t-observed formula
- All of the above
- What is statistical power?
- The ability for a statistic to produce a large effect size
- The power of a dataset to produce generalizable results
- The probability that a false null hypothesis will be rejected in a study
- The frequency that the critical value is on the right side of the sampling distribution
- High statistical power reduces the likelihood of a
- Type I error
- Type II error
- Rejecting the null hypothesis
- Retaining the null hypothesis
- Which of the following will raise statistical power?
- Increase the sample size
- Decrease the α value
- Using an unpaired two-sample t-test instead of a paired samples t-test size
- None of the above
- Studies with larger effect sizes tend to have higher statistical power. But why doesn’t this impact study design?
- Designs with larger effect sizes are harder to carry out than designs with smaller effect sizes
- Larger effect sizes usually require smaller α values
- The effect size is usually not under the researcher’s control
- All of the above
- When a study has low statistical power and the null hypothesis is rejected
- A Type II error has occurred
- The likelihood that the null hypothesis really is false is high
- The results are untrustworthy and seem unstable
- All of the above
- What is the relationship between group size and statistical power?
- Studies with equally sized groups usually have more power
- Studies with smaller groups tend to have more power
- A combined group size of 100 or more is necessary for a study to have high statistical power
- All of the above
- The correct interpretation of a p-value is the
- Probability that the researcher has committed a Type I error
- Likelihood that the results of a study would replicate in a future sample drawn from the population
- Likelihood that the results of a study are important and relevant to the researcher’s hypothesis
- Probability that a random sample would produce similar results, assuming the null hypothesis were perfectly true
- A p-value indicates the probability that
- The null hypothesis is false, given the results of the statistical analyses
- The strength of the effect size in the study
- The results of the statistical test are due to chance alone
- None of the above
- The homogeneity of variance assumption means that
- The sample’s standard deviation is the square root of the sample’s variance
- The two groups have similar levels of variability
- The dependent variable scores are interval- or ratio-level data
- All of the above
- The homogeneity of variance assumption is met when
- The sample’s standard deviation is the square root of the sample’s variance
- The dependent variable’s scores meet the requirements for interval- or ratio-level data
- The ratio of the two groups’ variance does not exceed 10:1
- All of the above
- One assumption of unpaired two-sample t-tests is that the sample sizes for the two groups are approximately equal. To meet this assumption, the larger group must not have a sample size more than _________ times larger than the smaller group
- 1
- 2
- 4
- 10
- If the relationship between the two variables in an unpaired two-sample t-test is very strong
- The two groups’ distributions will have little overlap
- The two groups’ distributions will overlap almost completely
- The α value will be high
- The t-critical value will be outside the tail of the distribution
- As the relationship between variables in an unpaired two-sample t-test decreases,
- The two groups’ distributions overlap more
- The overlap between the two groups will decrease
- The α value decreases
- The t-critical value moves towards the tail of the distribution
- Explain why the null hypothesis of in a paired-sample t-test is not appropriate for an unpaired samples t-test.
- Because calculating D values requires pairing scores across groups, and there is no logical way to pair data from independent groups.
- If a researcher has paired data, it is mathematically possible to still conduct an unpaired samples t-test. But why is this inadvisable?
- Because an unpaired samples t-test has lower statistical power than a paired samples t-test.
- Why is it that an extreme t-observed value and a very small p-value do not prove a null hypothesis true?
- Because it is never fully known whether the null hypothesis is true or not. Even if the null hypothesis is very easily rejected, it is still possible that the results are a Type I error. Rejecting the null hypothesis once in a single study does not conclusively disprove the null hypothesis forever.
- A student said, “When I rejected the null hypothesis, I obtained a p-value of .035. This means that there is only a 3.5% probability that the null hypothesis is true.” What is the problem with the student’s logic?
- A p-value is the probability that a random sample would produce the observed results, given the assumption that the null hypothesis was perfectly true. The student has the logic reversed: p does not state the probability that the null hypothesis is true, given the observed data.
- Why is the homogeneity of variance assumption important for an unpaired samples t-test?
- Because an unpaired samples t-test only tests whether two sample means are equal. But if two distributions have very different variances, the two samples could be very different while still having the same mean.
Chapter 11
- Instead of conducting an ANOVA, a researcher chooses to conduct a series of unpaired two-sample t-test. This increases which kind of Type I error?
- Testwise Type I error
- Experimentwise Type I error
- Aggregate Type I error
- Individual Type I error
- As the number of groups in an independent variable increases, the number of unpaired two-sample t-tests
- Reaches a plateau
- Increases faster
- Increases at the same rate
- Changes in an unpredictable fashion
- How does one perform a Bonferroni correction?
- Use a post hoc test that combines the data from some groups
- Expand the null hypothesis to accommodate multiple group means
- Divide the α by the number of tests
- None of the above
- Exploratory research
- Does not have an effect size
- Is not driven by pre-existing hypotheses
- Is not appropriate for the first research study on a topic
- Requires conducting an ANOVA
- The sampling distribution in an ANOVA is called a(n)
- ANOVA distribution
- Normal distribution
- t-distribution
- F-distribution
- The alternative hypothesis in an ANOVA is always
- That any two group means are not equal
- Appropriate for a one-tailed test
- The same for any ANOVA
- All of the above
- ____________________ created the ANOVA
- Karl Pearson
- Sir Ronald Fisher
- George W. Snedecor
- William Gosset
- In the ANOVA information table, the values are
- Predicted dependent variable values
- Calculated for each group and for the total sample
- Dependent variables averages
- All of the above
- The for all groups combined is the
- Predicted dependent variable when sample members’ group membership is unknown
- Grand mean
- Helpful for interpreting an ANOVA’s effect size
- All of the above
- In an ANOVA, the observed value is obtained from the
- Table of critical values
- Information table
- ANOVA table
- Post hoc test table
- Every number in the ANOVA table will always be
- Close to the sample average
- Whole numbers
- Zero or positive
- All of the above
- When the null hypothesis of an ANOVA is rejected, the observed value is
- Less than the critical value
- Greater than 1.96
- Greater than the critical value
- Greater than the α value
- The effect size for ANOVA is
- η2
- Glass’s Δ
- Cohen’s d
- Standardized mean difference
- The range of possible values for the effect size for an ANOVA is
- -∞ to +∞
- -3 to +3
- 0 to 1
- 0 to 10
- The interpretation of the effect size for an ANOVA is
- The improvement of predictions when compared to the grand mean
- The proportion of dependent variable variance shared with the independent variable
- The degree to which the squared residuals are reduced when using group membership to predict dependent variable values
- All of the above
- Some people like to interpret effect sizes by labeling them as “small,” “medium,” or “large.” What is the problem with this tendency?
- The same standards for “small,” “medium,” or “large” effect sizes may not be appropriate for all topics of study
- These labels do not correspond to the importance of an effect size
- The creator of these labels said that they were not appropriate for all research situations
- All of the above
- Which of the following is not an important component of effect size interpretation?
- Prior research
- Group standard deviations
- Real-life applicability
- Context
- In an ANOVA, the independent variable is _________________ data, and the dependent variable is _________________ data.
- Nominal-level, nominal-level
- Interval- or ratio level, interval- or ratio-level
- Nominal-level, interval- or ratio-level
- Ratio-level, interval- or ratio-level
- It is unnecessary to conduct a post hoc test when the
- Null hypothesis is rejected
- Null hypothesis is retained
- Researcher conducts a one-tailed test
- Effect size is large
- The purpose of a post hoc test is to determine why the ANOVA
- Sampling distribution is not normal
- Information table cannot be completed
- Null hypothesis was rejected
- Alpha value needs to be adjusted
- The harmonic mean is a measure of central tendency that
- Is equal to the 5% trimmed mean
- Compensates for the shortcomings of an ANOVA
- Can be calculated with ordinal-level data
- Gives disproportionate weight to smaller sample sizes
- Instead of examining every possible group difference (as in a Tukey’s post hoc test), some post hoc tests examine only selected group differences based on theory. These post hoc tests are called
- Cohen’s contrasts
- Planned contrasts
- A priori contrasts
- Explanatory contrasts
- Which post hoc test is most appropriate when there is a clear baseline group for other groups to be compared to?
- Tukey’s post hoc test
- Bonferroni-Dunn post hoc test
- Scheffé’s post hoc test
- Dunnett’s test
- Which is the most common post hoc test in the social sciences?
- Tukey’s post hoc test
- Bonferroni-Dunn post hoc test
- Scheffé’s post hoc test
- Dunnett’s test
- _______________ measures accuracy of predictions at the individual level.
- η2
- A Bonferroni correction
- The residual
- Tukey’s post hoc test
- _______________ measures accuracy of predictions at the overall level.
- η2
- A Bonferroni correction
- The residual
- Tukey’s post hoc test
- Which value will always be largest?
- SOSTotal
- SOSWithin
- SOSBetween
- None of these is consistently larger than the others
- Which value will always be largest?
- dfTotal
- dfWithin
- dfBetween
- None of these is consistently larger than the others
- What is another term for the residual?
- Statistical power
- Shared dependent variable variance
- Independent variable condition
- Error
- What are the two disadvantages to conducting a Bonferroni correction?
- (1) The correction is too severe and can lower alpha values too much, and (2) it does not address the problem of conducting a large number of unpaired samples t-tests.
- In ANOVA groups, what is the best value for predicting a sample member’s dependent variable score?
- The mean for the group that the individual belongs to.
- Explain the relationship between an unpaired samples t-test and an ANOVA in the context of the General Linear Model.
- The unpaired two-sample t-test is a special case of an ANOVA (i.e., the ANOVA is a generalization of the unpaired two-sample t-test). This means that if a person conducts an ANOVA with two groups, the results will be the exact same as the results of an unpaired two-sample t-test.
- What is the null hypothesis of an ANOVA, and what does it mean when that hypothesis is rejected?
- The null hypothesis for an ANOVA is that all group means are equal to one another (i.e., ). When this null hypothesis is rejected, it means that at least one group mean differs from at least one other group mean.
- How is the sampling distribution in an ANOVA different from the sampling distribution from a z-test and every type of t-test?
- The sampling distribution in an ANOVA (i.e., the F-distribution) is always positively skewed, whereas the z-distribution and t-distribution are symmetrical/unskewed.
Chapter 12
- When two variables are correlated, it indicates that the two variables are
- Interval-level data
- Related
- Continuous
- Predictable
- The two components of Pearson’s r to interpret are the
- Level of data and number
- Sign and level of data
- Sign and number
- Direct and inverse correlations
- A positive correlation is also called a(n)
- Direct correlation
- Nonlinear correlation
- Linear correlation
- Inverse correlation
- A negative correlation is also called a(n)
- Direct correlation
- Nonlinear correlation
- Linear correlation
- Inverse correlation
- What are the range of possible values for Pearson’s r?
- -∞ to +∞
- -3 to +3
- -1 to +1
- 0 to +1
- Which Pearson’s r value indicates the strongest correlation?
- +.42
- -.28
- -.70
- +.55
- Which Pearson’s r value indicates the weakest correlation?
- +.42
- -.28
- -.70
- +.55
- Which Pearson’s r value would indicate a dataset that has very few exceptions in its relationship between variables?
- -.52
- -.22
- +.05
- +.83
- Which Pearson’s r value would indicate a subtle and extremely weak relationship between its variables?
- -.52
- -.22
- +.05
- +.83
- The Pearson’s r equation automatically converts all the data into
- z-scores
- Standard deviations
- Covariances
- None of the above
- Which correlation will produce a scatterplot that consists of a series of dots on a straight line that extends from the upper left quadrant to the bottom right quadrant of the scatterplot?
- -1.0
- 0.0
- +0.5
- +1.0
- Which correlation will produce a scatterplot that consists of a series of dots that are scattered more-or-less randomly throughout the scatterplot?
- -1.0
- 0.0
- +0.5
- +1.0
- When the correlation is negative, sample members with _____________ scores on the independent variable tend to have _____________ scores on the dependent variable.
- High, high
- Low, high
- Average, average
- Average, low
- When the correlation is positive, sample members with _____________ scores on the independent variable tend to have _____________ scores on the dependent variable.
- High, high
- Low, high
- Average, average
- Average, low
- When a correlation is _________________ the dots in a scatterplot tend to be spread apart, while in a ____________________ correlation the dots in a scatterplot tend to be close together.
- Strong, weak
- Positive, negative
- Weak, strong
- Negative, positive
- The null hypothesis for a correlation coefficient is
- r = 0
- None of the above
- Which of the following is the alternative hypothesis for a two-tailed null hypothesis test of Pearson’s r?
- r = 0
- r ≠ 0
- r < 0
- r > 0
- Which of the following is the alternative hypothesis for a one-tailed null hypothesis test of Pearson’s r where the rejection region is completely on the left side of the sampling distribution?
- r = 0
- r ≠ 0
- r < 0
- r > 0
- Which of the following is the alternative hypothesis for a one-tailed null hypothesis test of Pearson’s r where the rejection region is completely on the right side of the sampling distribution?
- r = 0
- r ≠ 0
- r < 0
- r > 0
- To determine that the null hypothesis for a test of Pearson’s r should be rejected, a researcher should
- Expand the rejection region until it includes the critical value
- Follow the same steps as for any other null hypothesis test
- Ensure that the sampling distribution is non-normal and centered on zero
- Calculate the Pearson’s r first and then use that information to determine whether to conduct a one- or two-tailed test
- The effect size r2 is equal to
- Pearson’s r
- η2
- Glass’s Δ
- Cohen’s d
- One interpretation of r2 is to compare the improvement of predictions when compared to the
- Dataset’s z-scores
- Residuals calculated from the correlation coefficient
- Sample mean
- Baseline model without predictors
- According to Schönbrodt and Peruigini (2013), a sample size of approximately __________ is needed to produce stable correlations.
- 25
- 50
- 100
- 250
- When two variables are correlated, it is because
- Changes in the independent variable cause changes in the dependent variable
- Changes in the dependent variable cause changes in the independent variable
- Changes in a third variable cause changes in both the independent and dependent variables
- All of the above are possible reasons why variables are correlated
- If two variables are correlated, why is it impossible to state that changing one variable score will result in a change in the other variable’s score?
- Interventions or treatments may be very complex to design and implement
- The two variables may be correlated simply because they both relate to a third variable.
- Dependent variables are often not sensitive to changes from other variables
- All of the above
- Which of the following is an example of a positive correlation?
- Adolescents who earn higher grades use fewer drugs
- People who use more social media are better informed of news events
- Higher income countries have lower illiteracy rates
- All of the above
- Which of the following is an example of a negative correlation?
- As individuals watch more television, they read less
- Children with higher reading skills often have higher reading skills
- Larger cities usually have higher levels of pollution
- Prisoners who break fewer rules in prison usually commit fewer crimes after being released
- Which of the following is an example of a negative correlation?
- Younger adults tend to have fewer health problems
- Families with higher incomes generally eat at restaurants more often
- There is no relationship between people’s height and their political beliefs
- Countries with higher immunization rates tend to have lower disease rates
- Which of the following is an example of a negative correlation?
- Children who are more involved with sports are more physically fit
- Adults who vote more often are more involved with community affairs
- Individuals who are low in neuroticism tend to stay in their jobs longer
- People with more friends tend to have better coping skills
- Which of the following is an example of a positive correlation?
- Children who play more video games have shorter reaction times
- Schools where teachers are better trained have fewer disciplinary problems
- As the temperature increases, more people purchase ice cream
- Doctors who prescribe more antibiotics have fewer patients who need additional treatment
- Which of the following is an example of a positive correlation?
- Individuals with lower self-control often have fewer friends
- Smaller universities generally have students who are more involved with activities
- People who travel less usually feel more comfortable at home than people who travel more
- Students who send text messages frequently in class generally earn lower grades
- Give an example of two uncorrelated variables.
- Responses will vary, but the student’s response should be too variables that are unrelated (e.g., eye color and height).
- Explain the meaning of the phrase “correlation is not causation.”
- The phrase indicates that the existence of a correlation does not—by itself—prove that changes in one variable cause changes in the other.
- Under the baseline model for a correlation, what is every sample member’s predicted score on the dependent variable?
- The grand mean for the dependent variable (i.e., the mean of all dependent variable scores in the data).
- What happens to the pattern of dots on a scatterplot as the correlation between two variables strengthens?
- The dots group closer together.
Chapter 13
- The standardized regression equation is appropriate for situations where
- The independent and dependent variables are both interval- or ratio-level data
- The independent and dependent variables are both z-scores
- The independent variable is nominal-level data and the dependent variable is interval-level data
- An ANOVA requires a post hoc test
- The standardized and unstandardized regression equations are both used to
- Convert data into z-scores
- Link regression to the other procedures in the General Linear Model
- Make predictions about the dependent variable
- Enlarge the residuals of each sample member
- The regression line
- Is created via the regression equations
- Is also called the line of best fit
- Minimizes the vertical distance between each point and the regression line
- All of the above
- All regression predictions of dependent variable scores are
- On the regression line
- Larger than the actual dependent variable scores
- Larger than the independent variable scores
- All of the above
- It takes a minimum of ______________ points to draw a regression line
- One
- Two
- Three
- Four
- Converting all the data to the z-scores
- Strengthens the correlation
- Weakens the correlation
- Does nothing to the pattern of data in the scatterplot
- Converts the data to ratio-level scores
- The regression line will always pass through
- (10, 10)
- (0, 1)
- (, )
- (-1, +1)
- In the standardized regression equation, the y-intercept is always
- 0
- 1
- 5
- 10
- A positive residual indicates
- The person scored higher than predicted on the dependent variable
- The correlation is positive
- The person scored lower than predicted on the dependent variable
- The correlation is negative
- A negative residual indicates
- The person scored higher than predicted on the dependent variable
- The correlation is positive
- The person scored lower than predicted on the dependent variable
- The correlation is negative
- The vertical distance between a person’s point on the scatterplot and the regression line is the
- Predicted dependent variable value
- Residual
- z-score
- Standard deviation
- In the baseline model, every sample member is predicted to have a score equal to
- Their actual dependent variable score
- Their z-score
- The dependent variable standard deviation
- The dependent variable mean
- The regression line and the line for the baseline model intersect at
- (-1, +1)
- (, )
- (-3, +3)
- (-2, +2)
- The r2 is
- The degree of improvement in the predictions when compared to the baseline model
- The percentage decrease in the sum of the squared residuals when calculated the regression line compared to the sum of the squared residuals calculated from the baseline model
- The square of the correlation coefficient
- All of the above
- An r2 value of 0 indicates that
- Models based on the regression line are as good as predictions from the baseline model
- The residuals for the regression line are all zero
- The baseline model is a better tool for making predictions than the regression line
- The sum of the residuals using the regression line is zero
- An r2 value of 1 indicates that
- The residuals for the data add to 1
- The baseline model is a line that has a slope of +1
- All the predictions are perfectly accurate
- All of the above
- Because of regression towards the mean
- Over time the average score in a population will decline
- Outliers are difficult to predict
- The mean of a variable will increase when using a regression equation
- All of the above
- Regression towards the mean will be greatest for
- Outliers
- Scores close to the average
- Correlations of r > .5
- z-scores
- Regression towards the mean will be zero for
- Outliers
- Correlations of r > .5
- Subjects with independent variable scores that are equal to the mean
- z-scores
- Regression towards the mean was discovered by
- Karl Pearson
- Francis Galton
- William Gosset
- John Tukey
- Which would be a real-life situation where regression towards the mean would apply?
- A sports team that wins a championship is not as good of a team the following year
- After writing her best novel yet, an author’s following novel is not as good
- The worst salesperson in a company for a month improves in their performance the following month
- All of the above
- Graphing the regression line requires knowing at least ________ points that fall on the line.
- 2
- 3
- 5
- 10
- Which of the following is not an assumption of regression?
- A linear relationship between variables
- Interval- or ratio-level data for both variables
- Homogeneity of residuals
- All of the above
- What is the usual impact of restriction of range on a correlation coefficient?
- Its strength is reduced and the value is closer to zero
- It increase in strength and is closer to -1.0
- It increases in strength and is closer to +1.0
- It switches from positive to negative (or from negative to positive)
- Which of the following is an example that could cause restriction of range?
- A sample of English speakers is not representative of the population of people worldwide
- Requiring a minimum test scores for college admission and then finding the correlation between test scores and college GPA
- The outliers in a sample have the largest influence over the correlation value
- All of the above
- To determine whether outliers are distorting the relationship, the researcher should ______________________ and determine whether it changes the correlation.
- Re-calculate the correlation without the outliers
- Convert the data to z-scores
- Find the regression line
- Divide the sample in half
- If the data in a correlation coefficient consists of two or more groups,
- The data should be combined into one group before calculating a correlation coefficient or performing regression
- The groups should each occupy a different portion of the scatterplot
- The correlation for each subgroup should be similar to the correlation for the entire dataset
- All of the above
- As a member of the General Linear Model, the point-biserial correlation is a connection between Pearson’s r and the
- z-test
- one-sample t-test
- unpaired samples t-test
- ANOVA
- The slope of the standardized regression line is equal to what value?
- -1
- 0
- 1
- r
- Describe the appearance on the baseline model on a scatterplot.
- On a scatterplot, the baseline model is a horizontal line that at the mean for the dependent variable.
- Explain what homogeneity of residuals means and why it is important.
- Homogeneity of residuals is an assumption of correlation and regression which states that the dots in a scatterplot form an oval or circle (instead of a trumpet or a funnel). It is important because if the assumption is violated, it indicates that the regression line is a model that fits some of the data better than other parts of the data.
- Draw an example of a nonlinear relationship.
- Student responses will vary, but a correct response will be a scatterplot where the dots cannot be summarized as a straight line. They may instead for a parabola, or possibly will be a pattern of dots that change direction (e.g., by leveling off after an increase or after a decrease).
- Explain regression towards the mean.
- Regression towards the mean is a phenomenon where—on average—predicted dependent variable z-scores will be closer to the mean than the same individuals’ independent variable z-scores were.
- Also acceptable: Regression towards the mean occurs when researchers predict that individuals will be more average on the dependent variable than they were on the independent variable.
- What is the homogeneity of subgroups assumption, and what happens when it is violated?
- The homogeneity of subgroups assumption is that any subgroups within the sample have similar correlations between the independent and dependent variables. When this assumption is violated, then the correlation for the entire dataset will not be similar to the correlations calculated using each subgroup’s data
Chapter 14
- A one-variable chi-squared test most closely resembles which other member of the General Linear Model?
- z-test
- one-sample t-test
- Pearson’s r
- ANOVA
- The one-variable chi-square test has the null hypothesis that the independent variable groups’ data are equal to
- Theorized proportions of the dependent variable categories
- Mean scores on the dependent variable
- Standard deviations on the dependent variable
- z-scores of the dependent variable
- A chi-squared test is always a
- Test of residual variance
- Test that requires an α value of .05
- One-tailed test
- Two-tailed test
- The chi-squared sampling distribution is
- Normally distributed
- Negatively skewed
- Similar in shape to a t-distribution
- Positively skewed
- For a one-variable chi-squared test, the chi-squared observed equation produces a series of fractions that are added together. The number of fractions will always be equal to
- The number of groups in the dependent variable
- The number of groups in the independent variable
- n - 1
- n / 2
- The possible range of chi-squared observed values is
- -∞ to +∞
- -3 to +3
- 0 to 10
- 0 to +∞
- One criticism of null hypothesis statistical significance (NHST) testing procedures is that
- The p-value is continuous, yet NHST produces a dichotomous decision
- It can encourage researchers to tweak their study to make it artificially easier to reject the null hypothesis
- The difference between “significant” and “non-significant” results may be very subtle
- All of the above
- Publication bias
- Occurs when studies that reject the null hypothesis are more likely to get published
- Fixes the file drawer problem that plagues the social sciences
- Is the result of logical decisions about the priorities of scientists
- All of the above
- Publication bias is a result of problematic decisions from
- Journal editors
- Researches
- Peer reviewers
- All of the above
- When an odds ratio is equal to 1.0
- The two groups’ means are one standard deviation apart
- The effect size is statistically significant
- The observed and expected values are equal
- The groups are all the same size
- In a one-variable chi-squared test, if the odds ratio is 2.5, it indicates that
- There are 2.5 times more independent variable groups than were theorized
- There are 2.5 times more sample members in a group than were theorized
- There are 2.5 times fewer independent variable groups than were theorized
- There are 2.5 times fewer sample members in a group than were theorized
- In a one-variable chi-squared test, if the odds ratio is 4.1, it indicates that
- There are 4.1 times more independent variable groups than were theorized
- There are 4.1 times more sample members in a group than were theorized
- There are 4.1 times fewer independent variable groups than were theorized
- There are 4.1 times fewer sample members in a group than were theorized
- In a one-variable chi-squared test, if the odds ratio is 0.5, it indicates that
- There are 0.5 times as many independent variable groups than were theorized
- There are 0.5 times as many sample members in a group than were theorized
- There are 0.5 times as many dependent variable groups than were theorized
- The overall sample size is 0.5 times as large as was theorized
- When an odds ratio is precisely equal to 1.0
- The null hypothesis is a perfect model for the data
- The null hypothesis has been rejected
- The smallest group in the data has the highest mean
- All of the above
- If a one-variable chi-squared test retains the null hypothesis, it indicates that
- The results are statistically significant
- The proportion of the sample in each group is similar to the theorized group proportions
- The p-value is less than the α value
- The observed value is inside the rejection region of the sampling distribution
- The null hypothesis in a two-variable chi-squared test is that
- Group sizes are all equal to one another
- The proportions of each group are equal to their theorized proportions
- The independent and dependent variables are correlated
- The independent and dependent variable groups are unrelated
- For a two-variable chi-squared test, the chi-squared observed equation produces a series of fractions that are added together. The number of fractions will always be equal to
- The number of groups in the dependent variable
- The number of groups in the independent variable
- The number of cells in the table
- The sum of the number of independent and dependent variable groups added together
- The expected values in the rows of a two-variable chi-squared table will always sum up to
- The sum of the observed values in the same row
- The sum of the observed values in the columns
- The number of sample members divided by the number of cells in the row
- The sum of the number of independent and dependent variable groups added together
- The chi-squared sampling distribution is a theoretical distribution of
- Sample means
- Sample variances
- Sample standard deviations
- Differences between means
- Calculating an odds ratio for a two-variable chi-squared test requires choosing a
- Baseline group
- Outcome of interest
- Non-baseline group
- All of the above
- If the baseline and non-baseline groups are reversed, the odds ratio changes to the
- Original odds ratio multiplied by negative 1
- Original odds ratio multiplied by the sample size
- Reciprocal of the original odds ratio
- Original odds ratio divided by the number of independent variable groups
- In a two-variable chi-squared test, if the odds ratio is 1.0, it indicates that
- The two nominal variables are unrelated
- The null hypothesis needs to be retained
- The outcomes are equally likely for the baseline and the non-baseline groups
- All of the above
- In a two-variable chi-squared test, if the odds ratio is less than 1, it indicates that
- The outcome of interest is more likely for the baseline group
- The outcome of interest is more likely for the non-baseline group
- The baseline outcome is less likely for the non-baseline group
- The outcome of interest is equally likely for both groups
- In a two-variable chi-squared test, if the odds ratio is greater than 1, it indicates that
- The outcome of interest is more likely for the baseline group
- The outcome of interest is more likely for the non-baseline group
- The baseline outcome is more likely for the non-baseline group
- The outcome of interest is equally likely for both groups
- An odds ratio can be converted into ____________________ for every combination of groups and outcomes.
- The null hypothesis
- A z-score
- A predicted probability
- A proportion
- Meta-analysis
- Is possible because effect sizes can be averaged together
- Overcomes this subjectivity inherent in literature reviews
- Can resolve contradictions in the scientific literature
- All of the above
- When there are more than two groups in the independent or dependent variable in a two-variable chi-squared test,
- The researcher may need to calculate more than one odds ratio
- Multiple groups must serve as the baseline group
- The equation for an odds ratio becomes much more complex
- None of the above
- The relative risk describes the
- Odds ratio for z-scores
- Probability of the outcome of interest occurring for the non-baseline group compared to the baseline group
- Probability that the null hypothesis will be rejected in a two-variable chi-squared test
- Results of a meta-analysis for the dependent variable
- What does a relative risk of 1.0 indicate?
- The outcome of interest is more likely for the baseline group
- Equal proportions of the two groups experience the outcome of interest
- The outcome of interest is more likely for the non-baseline group
- The odds ratio would be a more appropriate effect size for the data
- What does a relative risk of value that was less than 1.0 indicate?
- The outcome of interest is more likely for the baseline group
- Equal proportions of the two groups experience the outcome of interest
- The outcome of interest is more likely for the non-baseline group
- The odds ratio would be a more appropriate effect size for the data
- What does a relative risk of value that was more than 1.0 indicate?
- The outcome of interest is more likely for the baseline group
- Equal proportions of the two groups experience the outcome of interest
- The outcome of interest is more likely for the non-baseline group
- The odds ratio would be a more appropriate effect size for the data
- A correlation coefficient for nominal data is called
- Pearson’s r
- The relative risk
- The odds ratio
- Phi (ϕ)
- Given an odds ratio of 4.0, what would the odds ratio be if the baseline and non-baseline groups were switched?
- 0.10
- 0.25
- 0.40
- 1.40
- Given an odds ratio of 0.33, what would the odds ratio be if the baseline outcome and the outcome of interest were switched?
- 1.00
- 1.33
- 3.00
- 33.00
- What is a problem with the effect size Cramér’s V?
- It can only have a negative value
- It requires interval- or ratio-level data to calculate
- When the null hypothesis is rejected, the value of V will always be zero
- Different numbers of categories in the variables can inflate or depress the V value
- State a consequence of publication bias.
- Publication bias distorts the literature and (usually) inflates effect sizes.
- Also acceptable: it encourages researchers to engage in p-hacking to tweak their studies to produce a p-value below .05.
- Also acceptable: it prevents readers from having access to the full body of research on a topic.
- Explain the similarities between the two-variable chi-squared test and Pearson’s r in the context of the General Linear Model.
- A two-variable chi-squared test is just a Pearson’s r correlation for nominal data. Also acceptable: the null hypothesis for both procedures is that there is no relationship between independent and dependent variables.
- What is the purpose of performing a meta-analysis?
- A meta-analysis combines results from several studies into one large study. This increases the sample size of the analysis beyond the sample size of any one study.
- Give an example of the type of data that would be appropriate for a two-variable chi-squared test?
- Responses will vary, but correct student responses will consist of two categorical (i.e., nominal or ordinal) variables.
- To calculate a chi-squared observed value, it is first necessary to find a series of expected counts. What is the meaning of these expected counts?
- Expected counts are the number of individuals that would be expected in the cell in a contingency table if the data perfectly matched the null hypothesis.
- What is the purpose of creating a contingency table in a chi-squared test?
- The contingency table displays all possible combinations of the independent and dependent variable values and the number of sample members in each group. This information can then be used to calculate expected counts, which are necessary to calculate the chi-squared observed value.
Chapter 15
- Which procedure(s) was/were designed to analyze multiple dependent variables simultaneously?
- Nonparametric statistics
- Multiple regression
- Logistic regression
- Multivariate statistics
- Which is the proper sequence for designing a research study?
- Formulate research questions 🡪 Choose methodology 🡪 Plan statistical analysis
- Create research survey 🡪 Formulate research questions 🡪 Plan statistical analysis
- Plan statistical analysis 🡪 Choose methodology 🡪 Formulate research questions
- Select sample 🡪 Plan statistical analysis 🡪 Formulate research questions
- Which procedure is appropriate when a dataset has two or more independent variables and one interval-level dependent variable?
- Multiple regression
- Logistic regression
- Kruskal-Wallis test
- Wilcoxon rank-sum test
- Comparing ______________________ can help determine which independent variable has the strongest relationship with the dependent variable.
- Logistic regression results
- Sample sizes
- β weights
- p-values
- The null hypothesis in multiple regression is
- r = 0
- b = 0
- α = 0
- The effect size for multiple regression is
- r2
- Cohen’s d
- The main effect
- R2
- An interaction occurs when
- The effect size in a multiple regression increases when independent variables are added to the data
- The impact of an independent variable varies according to the value of another independent variable
- The sample size increases in a study due to the recruitment of more subjects
- There is a relationship between the independent and dependent variables in a one-way ANOVA
- Multicollinearity occurs when
- Multiple regression fails to reject the null hypothesis
- The addition of an independent variable lowers a statistical model’s effect size
- Independent variables are highly correlated with one another
- The effect size is lowered because of the presence of an interaction
- A two-way ANOVA has
- Two independent variables
- An independent and dependent variable
- No ability to test for an interaction
- Cannot produce just one effect size value
- On a line graph, the presence of an interaction is indicated by
- Parallel lines
- Nonparallel lines
- Vertical lines
- None of the above
- When an interaction is present
- The relationship between the independent variables and the dependent variable is nuanced
- Results will not replicate in a future study
- The assumptions of multiple regression have been violated
- The effect size is not trustworthy
- Dummy coding is a way of preparing _____________ data for analysis in multiple regression
- Dependent variable
- Ordinal
- z-score
- Nominal
- Logistic regression requires
- z-score data
- Multiple dependent variables
- A nominal-level dependent variable
- All of the above
- The effect size for logistic regression and multiple logistic regression is
- η2
- The odds ratio
- Cohen’s d
- R2
- Simpson’s paradox occurs when
- Independent variables in a multiple regression are statistically significant, but the effect size is close to zero and indicates that the overall model is not statistically significant
- An independent variable’s β value changes its sign when another independent variable is added to the statistical model
- The results in a study contradict previous literature, yet are still correct
- A sample produces strong results for independent variables that are weakly correlated with the dependent variable
- A 3 x 4 ANOVA has ___________ independent variable(s) and ___________ dependent variable(s).
- 2, 1
- 3, 4
- 2, 12
- 12, 1
- Although nonparametric statistics have the advantage that they can be used for nominal and/or ordinal data, what is their disadvantage?
- It is impossible to convert interval-level data down to the level needed to conduct nonparametric statistics
- Nonparametric statistics cannot detect relationships between independent and dependent variables
- Nonparametric statistics have lower statistical power than parametric statistics
- All of the above
- Spearman’s ρ is a nonparametric statistic that most closely resembles
- ANOVA
- Tukey’s post hoc test
- The unpaired-samples t-test
- Pearson’s r
- What is the minimum level of data that Spearman’s ρ is appropriate for?
- Nominal
- Ordinal
- Interval
- Ratio
- The null hypothesis for the Wilcoxon rank-sum test is that
- The correlation between two nominal variables is zero
- The standard deviations of two groups are equal to one another
- Two groups’ medians are equal
- The distributions of ranks for two groups’ ordinal data are equal
- The AUC is the effect size for a Wilcoxon rank-sum test. The range of AUC values is
- -∞ to +∞
- -3 to +3
- -1 to +1
- 0 to +1
- What is the interpretation of the AUC effect size?
- The probability of randomly selecting a person in the first group who outscores a randomly selected person from the second group
- The probability that a person in the group of interest will experience the outcomes of interest
- The number of standard deviations between medians in the two groups
- None of the above
- The Kruskal-Wallis test most closely resembles
- ANOVA
- Tukey’s post hoc test
- The unpaired-samples t-test
- Pearson’s r
- Meta-analytic thinking can involve
- Reporting as much detailed statistical information as possible
- Provide enough information for a reader to replicate the study
- Considering the results of a study in the wider research context
- All of the above
- The alternative of conducting multivariate statistics is to
- Convert dependent variables into independent variables
- Use parametric statistics in the analysis
- Analyze each dependent variable separately
- All of the above
- A latent variable is a variable that is
- Created by summing together the scores on a series of dependent variables
- A theorized underlying variable that is the common source of the scores on multiple dependent variables
- Cannot be observed without distorting the data in the dependent variables
- All of the above
- Factor analysis, structural equation modeling, and MANOVA are all types of
- Multiple regression
- Nonparametric statistics
- Descriptive statistics
- Multivariate statistics
- What are the two data characteristics that determine the appropriate statistical analysis procedure?
- The number of independent and dependent variables and the level of data for each variable
- What is the interpretation of the R2 effect size?
- The R2 effect size is interpreted the same way as the r2 effect size. That is, it is a measure of the improvement of predictions when using the statistical model compared to the baseline model (or the percentage of dependent variable variance shared with the independent variable’s variance).
- Why is it important to investigate the presence of an interaction?
- Because sometimes the overall average impact of a treatment may not apply to a particular sample member—or any sample member at all. Interactions tell researchers whether an intervention will be more effective for some sample members than for others.
- Also acceptable: if a student makes the same point but in terms of the strength of a relationship between an independent variable and a dependent variable varying according to individuals’ scores on a second independent variable.
- What is the purpose of the Variance Inflation Factor (VIF) or Tolerance statistics?
- Both of these statistics test for the presence of multicollinearity among independent variables.
- Explain why Simpson’s paradox can be difficult to detect.
- Detecting Simpson’s paradox can be difficult because it requires including all relevant independent variables that contribute to the paradox in the statistical model. However, there is no way to know beforehand which variables may be relevant.
Document Information
Connected Book
Statistics for the Social Sciences 1st Edition Test Bank
By Russell T. Warne
Explore recommendations drawn directly from what you're reading
Statistics for the Social Sciences 1st Edition Test Bank
DOCX Ch. All in one Current
All chapters in this product are shown above