Professional Documents
Culture Documents
BUSINESS STATISTICS
Communicating with Numbers
Published by McGraw-Hill/Irwin, a business unit of The McGraw-Hill Companies, Inc., 1221 Avenue of the
Americas, New York, NY, 10020. Copyright © 2013 by The McGraw-Hill Companies, Inc. All rights reserved.
Printed in the United States of America. No part of this publication may be reproduced or distributed in any
form or by any means, or stored in a database or retrieval system, without the prior written consent of The
McGraw-Hill Companies, Inc., including, but not limited to, in any network or other electronic storage or
transmission, or broadcast for distance learning.
Some ancillaries, including electronic and print components, may not be available to customers outside the
United States.
1 2 3 4 5 6 7 8 9 0 RJE/RJE 1 0 9 8 7 6 5 4 3 2
ISBN 978-0-07-131761-0
MHID 0-07-131761-9
w w w .m hhe.com
A B OUT THE A U T H O RS
This is probably the best book I have seen in terms o f explaining concepts.
Brad McDonald, Northern Illinois University
The book is well written, more readable and interesting than most stats
texts, and effective in explaining concepts. The examples and cases are
p a rticu la rly good and effective teaching tools.
Andrew Koch, James Madison University
Key Features
Key to this text’s positive reception are six core features around which this text is built.
Integrated Introductory Cases. Realistic introductory cases that students can relate
to introduce each chapter topic and form the basis of several examples in the chapters.
Written as Taught. Topics are presented the way they are taught in class, beginning with
the intuition and explanation and concluding with the application.
Connect Business Statistics. Connect is an online system that gives students the tools
they need to be successful in the course. Through guided examples and LearnSmart
adaptive study tools, students receive guidance and practice to help them master the
topics.
I really like the case studies and the emphasis on writing. We are m aking a big
effort to incorporate more business w riting in our core courses so th a t meshes well.
Elizabeth Haran, Salem State University
I N T R O D U C T O R Y CAS E
In all o f these chapters, the opening case leads directly into the application questions that
students w ill have regarding the material. Having a strong and related case w ill certainly
provide more benefit to the student, as context leads to improved learning.
Alan Chow, University of South Alabama
This is an excellent approach. The student gradually gets the idea that he can look at a
problem— one which might be fairly complex— and break it down into root components.
He learns that a little bit o f math could go a long way, and even more math is even more
beneficial to evaluating the problem.
Dane Peterson, Missouri State University
and build skills to communicate results
WR I T I N G WI T H S T A T I S T I C S
turn it into a report that
will be useful to an
Callie Fitzpatrick, a research ana
lyst with an investment firm, has organization. I will strive
been asked to write a report sum
marizing the weekly stock perfor
mance of Home Depot and Lowe’s.
to have my students learn
Her manager is trying to decide
whether or not to include one of from these examples.
these stocks in a client’s portfolio
and the average stock performance Bruce P. Christensen,
is one o f the factors influencing
their decision. Callie decides to
use descriptive measures to sum
Weber State University
marize stock returns in her report,
as well as provide confidence in
tervals for the average return for
Home Depot and Lowe’s. She col This is an excellent
lects weekly returns for each firm
for the first eight months of 2010. approach. . . . The ability
A portion o f the return data is
shown in Table 8.4; the complete to translate numerical
dat
ava
Sample Grim news continues to distress the housing sector. On August 24, 2010, Reuters reported information into words that
that the sales of previously owned U.S. homes took a record plunge in July to the slow
Report— est pace in 15 years. Combine this fact with the continued fallout from the subprime others can understand is
mortgage debacle, a sluggish economy, and high unemployment, and the housing sector
Weekly Stock appears quite unstable. Have these unfavorable events managed to trickle down and harm critical.
the financial performance of Home Depot and Lowe’s, the two largest home improvement
Performance: retailers in the U.S.?
One way to analyze their financial stability is to observe their stock performance dur Scott Bailey, Troy University
Home Depot ing this period. In order to make valid statements concerning the reward of holding these
stocks, weekly return data for each firm were gathered from January through August of
vs. Lowe's 2010. Table 8.A summarizes the important descriptive statistics. Excellent. Students need to
T A B L E 8 . A D escriptive Statistics fo r W eekly Returns o f Hom e D e p o t and Lowe's (n = 34)
This is easy for students to follow and I do get the feeling ... the sections are spoken language.
Zhen Zhu, University of Central Oklahoma
that make the content more effective
FIGURE 2.9 C o n stru ctin g a h isto g ra m fro m raw data w ith Excel
C o nstruct th e 80% co nfid en ce in te rv a l o f th e p o p u la tio n a. Derive a 99% confidence interval of the average
m ean. time taken by a college student to solve a Sudoku
Applied exercises from Co n s tru c t th e 90% co nfid en ce in te rv a l o f th e p o p u la tio n puzzle.
b. What assumption is necessary to make this
The Wall Street Journal, Kiplinger's, happens to th e m a rg in o f e rro r as th e co nfid en ce inference?
Increases fro m 80% to 90%?
Fortune, The New York Times, USA 29. Executive co m p e n sa tio n has risen d ra m a tic a lly b e yo nd th e
rising levels o f an average w o rke r's w a ge over th e years.
Today, various websites ou n s Sarah is an MBA s tu d e n t w h o decides to use h e r statistica l
w e ig h t loss p ro gra m claim s th a t w ith its skills to e stim a te th e m ean CEO co m p en satio n in 2010 fo r
– Census.gov, Zillow.com, e n d e d h e a lth y d ie t re g im e n , users lose s ig n ifica n t a ll la rg e co m p an ie s in th e U n ite d States. She takes a ra n do m
with in a m o n th . In o rd e r to e stim a te th e m ean sam ple o f six CEO com p en satio n s.
Finance.yahoo.com, loss o f a ll custom ers, a n u tritio n is t takes a sam ple
Compensation
ESPN.com; and more. e te rs and records th e ir w e ig h t loss o ne m o n th a fte r
Firm (in $ millions)
lin in g th e p ro g ra m . He co m p ute s th e sam ple m ean and
th e stan d ard d e v ia tio n o f w e ig h t loss as 12.5 p o u nd s and Intel 8.20
9.2 pounds, respectively. He believes th a t w e ig h t loss is like ly Coca-Cola 2.76
to be n o rm a lly d istrib u te d . Wells Fargo 6.57
a. C alculate th e m a rg in o f erro r w ith 95% confidence. Caterpillar 3.88
b. C o m p ute th e 95% co nfid en ce in te rv a l o f th e p o p u la tio n McDonald's 6.56
m ean.
U.S. Bancorp 4.10
25. The m a n ag er o f The Cheesecake Factory in Boston re p orts
Source: http://finance.yahoo.com .
th a t on six ra n d o m ly selected weekdays, th e n u m b e r o f
cu sto m e rs served w as 1 2 0 , 1 3 0 , 90, 2 0 5 , 185, a nd 220. She a. H ow w ill Sarah use th e above in fo rm a tio n to p ro vid e a
fo llo w s a n orm al d is trib u tio n . C o nstruct a 90% co nfid en ce o f a ll large co m p an ie s in th e U n ited States?
in te rv a l o f th e average n u m b e r o f cu stom ers served on b. W hat a ssum ptio n d id Sarah m ake fo r d e riv in g th e
weekdays. in te rva l estim ate?
26. A ccordin g to a re ce nt survey, h ig h school g irls average c. H ow can Sarah reduce th e m a rg in o f e rro r re p o rte d in
Assum e th a t th e survey w as based on a ran do m sam ple 30. A p rice -e arnin g s ra tio o r P/E ra tio is ca lculate d as a firm 's
o f 36 h ig h school g irls . The sam ple standard d e v ia tio n is share p rice com pared to th e in com e o r p ro fit earned by
co m p u te d as 10 te x t messages daily. th e firm per share. G enerally, a h ig h P/E ra tio suggests th a t
a. C alculate th e m a rg in o f e rro r w ith 99% confidence. investors are e xp e ctin g h ig h e r e arnings g ro w th in th e fu tu re
b. W hat is th e 99% co nfid en ce in te rva l o f th e p o p u la tio n com pared to co m p an ie s w ith a lo w e r P/E ra tio . The fo llo w in g
m ean te x ts th a t a ll h ig h school g irls send daily? ta b le show s th e P/E ra tio s fo r a sam ple o f firm s in th e
Conceptual Review
At the end of each chapter, we provide a conceptual review that provides a more
holistic approach to reviewing the material. This section revisits the learning outcomes
and provides the most important definitions, interpretations, and formulas.
Conceptual Review
LO 6.1 Describe a continuous random variable.
A continuous random variable is c h a r a c t e r i z e d b y ( in f in ite ly ) u n c o u n t a b l e v a l
u e s a n d c a n t a k e o n a n y v a l u e w i t h i n a n in te r v a l. T h e p r o b a b i l i t y t h a t a c o n t i n u o u s
r a n d o m v a r ia b le X a s s u m e s a p a r t i c u l a r v a lu e x is z e r o , th a t is , P (X = x )= 0. T hus,
f o r a c o n t i n u o u s r a n d o m v a r i a b l e , w e c a l c u l a t e t h e p r o b a b i l i t y w i t h i n a s p e c if i e d
in te r v a l. M o r e o v e r , th e f o l l o w i n g e q u a l i t i e s h o ld : P (a ≤ X ≤ b ) = P (a < X < b) =
P (a < X < b ) = P (a < X < b ) .
The probability density function f(x) o f a continuous random variable X describes
the relative likelihood that X assum es a value within a given interval. T he probability
P(a ≤ X ≤ b)is the area u n d e rf (x ) betw een points a and b .
For any value x o f the random variable X, the cumulative distribution function F(x)
is defined as F(x) = P(X ≤ x).
Most texts basically list what one should have learned but don't add much to that.
You do a good job o f reminding the reader o f what was covered and what was most
important about it.
Andrew Koch, James Madison University
They have gone beyond the typical [summarizing formulas] and I like the
structure. This is a very strong feature o f this text.
Virginia M. Miori, St. Joseph's University
What technology connects students ...
McGraw-Hill Connect®
Business Sta tistics
McGraw-Hill Connect Business Statistics is an online assignment and assessment
solution that connects students with the tools and resources they’ll need to achieve
success through faster learning, higher retention, and more efficient studying. It
provides instructors with tools to quickly select content for assignments according
to the topics and learning objectives they want to emphasize.
Student Resource Library. The Connect Business Statistics Student Library is the
place for students to access additional resources. The Student Library provides quick
access to recorded lectures, practice materials, the eBooks, data files, PowerPoint files,
and more.
to success in business statistics?
Student Reporting. Connect Business Statistics keeps instructors informed about how
each student, section, and class is performing, allowing for more productive use of lecture
and office hours. The progress-tracking function enables you to:
• View scored work immediately and track
individual or group performance with
assignment and grade reports.
• Access an instant view of student or class
performance relative to topic and learning
objectives.
• Collect data and generate reports required by
many accreditation organizations, such as
AACSB.
Instructor Library. The Connect Business Statistics Instructor Library is your reposi
tory for additional resources to improve student engagement in and out of class. You
can select and use any asset that enhances your lecture. The Connect Business Statistics
Instructor Library includes:
• eBook
• PowerPoint presentations
• Test Bank
• Instructor’s Solutions Manual
• Digital Image Library
to success in business statistics?
McGraw-Hill
CONNECT® PLUS
BUSINESS STATISTICS
Connect®Plus Business Statistics includes a seamless integration of an eBook and
Connect Business Statistics, with rich functionality integrated into the product.
Tegrity Campus:
Lectures 24/7
Tegrity Campus is integrated in Connect to help make your class time available 24/7.
With Tegrity, you can capture each one of your lectures in a searchable format for stu
dents to review when they study and complete assignments using connect. With a simple
one-click start-and-stop process, you can capture everything that is presented to students
during your lecture from your computer, including audio. Students can replay any part
of any class with easy-to-use browser-based viewing on a PC or Mac.
Educators know that the more students can see, hear, and experience class resources,
the better they learn. In fact, studies prove it. With Tegrity Campus, students quickly
recall key moments by using Tegrity Campus's unique search feature. This search helps
students efficiently find what they need, when they need it, across an entire semester of
class recordings. Help turn all your students’ study time into learning moments immedi
ately supported by your lecture. To learn more about Tegrity, watch a two-minute Flash
demo at http://tegritycampus.mhhe.com.
What software is available w ith this text?
Access Card ISBN: 0077426274 Note: Best option fo r both Windows and Mac users.
Minitab®/SPSS®/JMP®
Minitab® Student Version 14 ISBN: 007305237X
These software tools are available to help students solve the business statistics exercises
in the text. Each can be packaged with any McGraw-Hill business statistics text. SPSS
SV18 is both Windows and Mac compatible. The software expires 13 months after
installation. JMP SV8 is also Windows and Mac compatible. It expires 12 months after
installation. The software is current as of the publication of this text.
The Online Learning Center (OLC) provides the instructor with a complete Instructor’s
Solutions Manual in Word format, the complete Test Bank in both Word files and computer
ized EZ Test Online format, Instructor PowerPoint slides, text art files, an introduction to
ALEKS®, an introduction to McGraw-Hill Connect®Business Statistics, and more.
All test bank questions are available in an EZ Test online. Included are a number of
multiple choice, true/false, and short-answer questions and problems. The answers to all
questions are given, along with a rating of the level of difficulty, chapter goal the question
tests, Bloom’s taxonomy question type, and the AACSB knowledge category.
CourseSmart
ISBN: 0077501381
CourseSmart is a convenient way to find and buy eTextbooks. CourseSmart has the
largest selection of eTextbooks available anywhere, offering thousands of the most
commonly adopted textbooks from a wide variety of higher-education publishers.
Course Smart eTextbooks are available in one standard online reader with full text
search, notes and highlighting, and e-mail tools for sharing notes between classmates.
Visit www.CourseSmart.com for more information on ordering.
ALEKS
ALEKS is an assessment and learning program that provides individualized instruction
in Business Statistics, Business Math, and Accounting. Available online in partnership
with McGraw-Hill/Irwin, ALEKS interacts with students much like a skilled human
tutor, with the ability to assess precisely a student’s knowledge and provide instruc
tion on the exact topics the student is most ready to learn. By providing topics to meet
individual students’ needs, allowing students to move between explanation and practice,
correcting and analyzing errors, and defining terms, ALEKS helps students to master
course content quickly and easily.
ALEKS also includes a new instructor module with powerful, assignment-driven
features and extensive content flexibility. ALEKS simplifies course management and
allows instructors to spend less time with administrative tasks and more time directing
student learning. To learn more about ALEKS, visit www.aleks.com.
Dear Colleague,
As te x tb o o k a uthors, and more importantly, as instructors of business statistics, w e recognize the
great importance placed on accuracy. With this in mind, we have taken the following steps to
1. We received detailed feedback from over 150 instructor reviews, starting with first draft
manuscript through the final draft submitted to the publisher. Each review contributed
in significant ways to the accuracy of the content.
2. We personally class-tested the manuscript with our students in various drafts, continually
3. Each of us wrote, reviewed, and carefully checked all of the end-of-chapter material.
4. A developm ental editor w ent through each sentence to ensure that our language was as
clear as possible.
5. Multiple accuracy checkers reviewed each chapter and its accompanying end-of-chapter
material— once w hen the final manuscript was submitted to the publisher, and again when
8. Our Solutions Manual and Test Bank were reviewed by multiple independent accuracy
checkers.
Given the steps taken above, w e have the utmost confidence that you and your students will have
CONTACT INFO R M A TIO N We would be grateful to hear from any and all users of this text and its
supplements. Send your comments and suggestions to sjaggia@calpoly.edu or akelly@suffolk.edu.
Sincerely,
AC KNOWLEDGMENTS
We could not have developed our approach to teaching business statistics in isolation.
Many thanks go to a multitude of people for helping us make our idea come to fruition.
First and foremost, we would like to thank our families and friends for their support,
feedback, and patience throughout what has been an arduous, but rewarding, project. We
are especially grateful to Chandrika, Minori, John, Megan, and Matthew for bearing with
us on a daily basis.
We thank Scott Isenberg for believing that we could write this text before a single
chapter was written; Fr. James Woods (Boston College) for providing a jump-start to
Chapter 1; David Chelton for excellent assistance as a content editor; and Chandrika
Jaggia for invaluable assistance with the initial format of the book. We are indebted
to Jerzy Kamburowski (University of Toledo) for his outstanding accuracy check and
his insistence on mathematical precision and consistency. Preliminary editions of our
book were used at Suffolk University, Boston College, C al Poly San Luis Obispo,
and the University of Kentucky, and we thank the many students and instructors who
provided comments and suggestions. Special thanks go to Timothy Lambie-Hanson,
Kossi Makpayo, and Sara Shorba. Thanks too to Minitab, Inc. for the use of their
software.
The editorial staff of McGraw-Hill/Irwin are deserving of our gratitude for their guid
ance throughout this project, especially Wanda Zeman, Steve Schuetz, Tim Vertovec,
Ann Torbert, Dean Karampelas, Christine Vaughan, Michael McCormick, Mary Kazak
Sander, Keri Johnson, and Daryl Horrocks.
Reviewers
The text has benefited immensely from reviewers’ helpful suggestions, keen insights, and
constructive criticisms. We are very grateful to the following professors for taking the
time to provide valuable feedback throughout the development process:
PART ONE
Introduction
CHAPTER 1 Statistics a n d D ata 2
PART T W O
Descriptive Statistics
CHAPTER 2 Tabular an d Graphical M eth o d s 16
CHAPTER 3 Numerical Descriptive M easures 52
PART THREE
Probability and Probability Distributions
CHAPTER 4 In tro d u c tio n to Probability 94
CHAPTER 5 D iscrete Probability D istributions 134
CHAPTER 6 C o n tin u o u s Probability D istributions 172
PART FOUR
Basic Inference
CHAPTER 7 S am pling an d S am pling D istributions 206
CHAPTER 8 Estimation 240
CHAPTER 9 H y p o t h e s i s Testin g 270
CHAPTER 10 Statistical I n f e re n c e C o n c e r n i n g Two P o p u l a t i o n s 304
CHAPTER 11 Statistical I n f e re n c e C o n c e r n i n g V ariance 334
CHAPTER 12 C h i-S q u a r e Tests 360
PA R T FIVE
Advanced Inference
CHAPTER 13 Analysis of V ariance 386
CHAPTER 14 R eg ress io n Analysis 422
CHAPTER 15 Inference with R egression M odels 456
CHAPTER 16 R e g re ss io n M o d e l s for N o n lin e a r R e la tio n sh ip s 490
CHAPTER 17 R e g re ss io n M o d e ls w ith D u m m y Variables 516
PART SIX
Supplementary Topics
CHAPTER 18 Tim e Series a n d F o re c a s ti n g 544
CHAPTER 19 R eturns, In dex N u m b e r s , a n d Inflation 584
CHAPTER 20 N o n p a r a m e t r i c Tests 608
APPENDIXES
APPENDIX A Tables 646
APPENDIX B A n sw e r s to E v e n - N u m b e r e d Exercises 659
Glossary 693
P h o t o Credits 701
I ndex 703
CONTENTS
2 .4 Scatterplots 41
Using Excel to C onstruct a S c a tte rp lo t 43
C alculating th e p th p e rcentile 61
Death w ith the M e th o d o f Losing Weight 128 Characteristics o f th e N orm al D is trib u tio n 178
The Standard N orm al Variable 179
C onceptual Review 129
Finding a P rob a b ility fo r a Given z Value 180
Additional Exercises an d Case Studies 130 Finding a z Value fo r a Given P rob a b ility 182
Case Studies 132 Revisiting th e Empirical Rule 184
The Discrete P ro b a b ility D is trib u tio n 137 The Lo gnorm al D is trib u tio n 196
Using Excel fo r th e Lognorm al D is trib u tio n 198
5 .2 Expected Value, Variance, an d S tan dard
Deviation 141 W riting w ith Statistics 200
Classic Case o f a "B ad” Sample: The Literary Digest Debacle Case Studies 268
o f 1936 208
S am pling M e thod s 209
CHAPTER 9 __
The Special Election to Fill Ted Kennedy's Senate Seat 211
7.2 The Sampling D istribution o f the Sample Mean 213 Hypothesis Testing 270
The Expected Value and th e Standard D eviatio n o f th e
Introductory Case: Undergraduate
Sam ple Mean 214
Study Habits 271
S am pling fro m a N orm al P opu la tion 215
The C entral L im it Theorem 216
9.1 Introduction to Hypothesis Testing 272
The Decision to "R eject" o r "N o t R eject" th e
7.3 The Sampling D istribution o f the Sample N ull H ypothesis 272
Proportion 219
D efinin g th e N ull H ypothesis and the
The Expected Value and th e Standard D eviatio n o f th e
A lte rn a tive H ypotheses 272
Sample P ro p o rtio n 219
Type I and Type II Errors 275
Synopsis o f Introductory Case 223
9.2 Hypothesis Test of the Population Mean When σ
7.4 The Finite Population Correction Factor 224 Is Known 277
7.5 Statistical Quality Control 226 The ρ -Value A pproach 278
C ontrol Charts fo r Q u a n tita tiv e Data 227 The C ritical Value A pproach 281
Using Excel to Create a C ontrol C hart 229 C onfid ence Intervals and Two-Tailed H ypothesis
Tests 285
C ontrol Charts fo r Q u a lita tive Data 229
Using Excel to Solve H ypothesis Tests 285
Writing w ith Statistics 233
O ne Last Remark 287
Sample Report— Customer Wait Time 233
9.3 Hypothesis Test of the Population Mean When σ
Conceptual Review 234 Is Unknown 288
Additional Exercises and Case Studies 236 Using Excel to C alculate ρ -Val ues fro m th e
tdt D is trib u tio n 290
Case Studies 238
CHAPTER 13________________
CHAPTER 11
Nominal versus Real Rates of Return 588 Writing with Statistics 638
Index 703
BUSINESS STATISTICS
C H
1
A P T E R
Statistics and Data
LEARNING OBJECTIVES
After reading LO 1.1 Describe the importance of statistics.
this chapter
you should LO 1.2 Differentiate between descriptive statistics and inferential statistics.
be able to:
LO 1.3 Explain the need for sampling and discuss various data types.
Every day we are bombarded with data and claims. The analysis of data and the predictions made
from data are part of the field of statistics. Virtually every phase of human activity incorporates sta
tistics, yet most people do not know they are using it. In this first chapter, we are going to introduce
some important terms that will help us describe different aspects of statistics and their practical im
portance. We will be using these terms freq uently throughout the rest of the book. You are probably
familiar with some of these ideas already, from reading or hearing about opinion polls, surveys, and
the all-pervasive product ads. Our goal in this chapter is to place what you already know about these
uses of statistics within a framework of terms and descriptions that we can then use for explaining
where they came from and what they really mean. A proper understanding of statistical ideas and
concepts can help you understand what politicians or advertisers are really saying, or not saying. But
it can also help you understand more of the real world around us, including business, finance, health,
social interactions— just about any area of contemporary human activity.
I N T R O D U C T O R Y CASE
Tween Survey
Luke McCaffrey ow ns a ski resort tw o hours ou tsid e Boston, Massachusetts, and is in need o f a new m a rke t
ing m anager. He is a fa irly to u g h in te rv ie w e r and believes th a t th e person in this p o sitio n should have a basic
u n d e rsta n d in g o f data fund a m e n tals, in c lu d in g som e backg ro un d w ith statistical m ethods. Luke is p a rticula rly
interested in serving th e needs o f th e "tw e e n " p o p u la tio n (children aged 8 to 12 years old). He believes th a t tw e en
sp e n ding p o w e r has gro w n over th e past fe w years, and he w ants th e ir skiing experience to be m em ora b le so th a t
th e y w a n t to return. A t th e end o f last year's ski season, Luke asked 20 tw eens fo u r specific questions.
Q1. On yo u r car drive to th e resort, w h ich radio sta tio n was playing?
Q2. On a scale o f 1 to 4, rate th e q u a lity o f th e food at the resort (where 1 is poor, 2 is fair, 3 is good, and 4 is excellent).
Q3. Presently, th e m ain d in in g area closes at 3:00 pm . W hat tim e d o you th in k it should close?
Q4. H ow m uch o f y o u r own m on e y d id you spend at th e lod g e today?
The responses to these questions are show n in Table 1.1; these data are also fo u n d on th e te x t w ebsite and are
labeled T w e en S u rv e y .
Example 2. A gambler predicts that his next roll of the dice will be a lucky 7
because he did not get that outcome on the last three rolls.
Problem with conclusion: As we will see later in the text when we discuss prob
ability, the probability of rolling a 7 stays constant with each roll of the dice. It
does not become more likely if it did not appear on the last roll or, in fact, any
number of preceding rolls.
Example 3. On January 10, 2010, nine days prior to a special election to fill the
U.S. Senate seat that was vacated due to the death of Ted Kennedy, a Boston
Globe poll gave the Democratic candidate Martha Coakley a 15-point lead
over the Republican candidate Scott Brown. On January 19, 2010, Brown won
52% of the vote compared to Coakley’s 47% and became a U.S. senator for
Massachusetts.
Problem with conclusion: Critics accused the Globe, which had endorsed Coakley,
of purposely running a bad poll to discourage voters from coming out for Brown.
In reality, by the time the Globe released the poll, it contained old information
from January 2-6, 2010. Even more problematic was that the poll included people
who said that they were unlikely to vote!
Example 4. Starbucks Corp., the world’s largest coffee-shop operator, reported that
sales at stores open at least a year climbed 4% at home and abroad in the quarter
ended December 27, 2009. Chief Financial Officer Troy Alstead said that “the
U.S. is back in a good track and the international business has similarly picked
u p .. . . Traffic is really coming back. It’s a good sign for what w e’re going to see
for the rest of the year” (http://www.bloomberg.com, January 20, 2010).
Problem with conclusion: In order to calculate same-store sales growth, which
compares how much each store in the chain is selling compared with a year ago,
we remove stores that have closed. Given that Starbucks closed more than 800
stores over the past few years to counter large sales declines, it is likely that the
sales increases in many of the stores were caused by traffic from nearby, recently
closed stores. In this case, same-store sales growth may overstate the overall
health of Starbucks.
Types of Data
Sample data are generally collected in one of two ways. Cross-sectional data refers
to data collected by recording a characteristic of many subjects at the same point in
time, or without regard to differences in time. Subjects might include individuals,
households, firms, industries, regions, and countries. The tween data presented in
Table 1.1 in the introductory case is an example o f cross-sectional data because
it contains tween responses to four questions at the end o f the ski season. It is
unlikely that all 20 tweens took the questionnaire at exactly the same time, but the
differences in time are o f no relevance in this example. Other examples o f cross-
sectional data include the recorded scores o f students in a class, the sale prices of
single-family homes sold last month, the current price of gasoline in different states
in the U.S., and the starting salaries o f recent business graduates from The Ohio State
University.
Time series data refers to data collected by recording a characteristic of a subject
over several time periods. Time series can include daily, weekly, monthly, quarterly, or
annual observations. Examples of time series data include the monthly sales of cars at a
dealership in 2010, the daily price of IBM stock in the first quarter of 2010, the weekly
exchange rate between the U.S. dollar and the euro, and the annual growth rate of India in
the last decade. Figure 1.1 shows a plot of the real (inflation-adjusted) GDP growth rate
of the United States from 1980 through 2010. The average growth rate for this period is
2.7%, yet the plot indicates a great deal of variability in the series. It exhibits a wavelike
movement, spiking downward in 2008 due to the economic recession before rebounding
in 2010.
Figure 1.1 Real GDP growth rate from 1980 through 2010
espn.com Professional and college teams'scores, rankings, standings, individual player statistics.
EXERCISES 1. 2
1. It came as a big surprise when Apple's touch screen iPhone a. Describe the relevant population and the sample.
4, considered by many to be the best sm artphone ever, was b. Do the numbers 22% and 17% represent the population
found to have a problem (The New York Times, June 2 4 , 2010). parameters or the sample statistics?
Users complained of weak reception, and sometimes even
6. Go t o h t t p : / / w w w . f i n a n c e . y a h o o . c o m / t o g e t a c u r r e n t sto c k
dropped calls, when they cradled th e phone in their hands in
q u o t e f o r G o o g le , Inc. (ticker s y m b o l = GOOG). T h e n , click
a particular way. A quick survey at a local store found that 2%
o n h isto rical p ric e s t o re c o rd t h e m o n t h l y a d j u s t e d clo se
of iPhone 4 users experienced this reception problem.
p rice o f G o o g l e s to c k in 2 0 1 0 . C r e a t e a t a b l e t h a t u s e s th is
a. Describe th e relevant population. inform ation. W h at ty p e of d a ta d o th e s e n u m b e r s rep re sen t?
b. Does 2% d enote th e population param eter or the C o m m e n t o n t h e data.
sample statistic?
7. Ask 20 of your friends w hether they live in a dormitory,
2. M any p e o p le reg ard v id e o g a m e s as a n o b se ss io n for a rental unit, or other form of accommodation. Also find
yo un g ste rs, b u t in fact, t h e a v e r a g e a g e of a v id eo g a m e player is out their approximate monthly lodging expenses. Create a
35 years (Reuters.com, A u g u s t 2 1 , 2009). Is t h e valu e 35 likely t h e table that uses this information. What type of data do these
actual o r t h e e s t im a te d a v e r a g e a g e o f t h e p o p u la tio n ? Explain. numbers represent? Comment on the data.
3. An accounting professor wants to know th e average GPA of 8. Go to http://www.zillow.com/ and find the sale price data
the students enrolled in her class. She looks up information of 20 single-family homes sold in Las Vegas, Nevada, in the
on Blackboard about th e students enrolled in her class and last 30 days. You must include in the data set the sale price,
computes th e average GPA as 3.29. the num ber of bedrooms, the square footage, and the age of
a. Describe th e relevant population. the house. What type of data do these numbers represent?
b. Does th e value 3.29 represent the population parameter Comment on the data.
or th e sample statistic? 9. T h e F ed eral R e se rv e B ank o f St. Louis is a g o o d s o u r c e
4. Business g r a d u a t e s in t h e U.S. w ith a m a r k e tin g c o n c e n tr a tio n fo r d o w n l o a d i n g e c o n o m i c d a t a . Go t o h t t p : / / r e s e a r c h
e a r n h ig h salaries. A ccording t o t h e B u reau o f L abor Statistics, . s t l o u i s f e d . o r g / f r e d 2 / t o e x t r a c t q u a r t e r l y d a t a o n g r o ss
t h e a v e r a g e a n n u a l salary for m a r k e t i n g m a n a g e r s w a s p r iv a te sa v in g (GPSAVE) f r o m 2 0 0 7 t o 2 0 1 0 (16 o b s e r v a t i o n s ) .
$1 04 ,4 00 in 2007. C r e a te a t a b l e t h a t u s e s th is i n f o r m a t i o n . P lo t t h e d a t a o v e r
b. Do you think th e average salary of $104,400 was 10. Another good source of data is the U.S. Census Bureau. Go
com puted from th e population? Explain. to http://www.census.gov/ and extract the most recent
state median income by family size (total) for Alabama,
5. Recent research suggests that depression significantly
Arizona, California, Florida, Georgia, Indiana, Iowa, Maine,
increases the risk of developing dementia later in life (BBC
Massachusetts, Minnesota, Mississippi, New Mexico,
News, July 6 , 2010). In a study involving 949 elderly persons,
North Dakota, and Washington. What type of data do
it was reported that 22% of those who had depression w ent
these numbers represent? Comment on the regional
on to develop dementia, compared to only 17% of those
differences in income.
who did not have depression.
In order to choose the appropriate statistical methods for summarizing and analyzing
data, we need to distinguish between different measurement scales. All data measure
ments can be classified into one of four major categories: nominal, ordinal, interval, and
ratio. Nominal and ordinal scales are used for qualitative variables, whereas interval and
ratio scales are used for quantitative variables. We discuss these scales in ascending order
of sophistication.
stock is traded on Nasdaq and the number 1 to show that a company’s stock is traded
on NYSE. In tabular form:
Excellent 4 Fair 2
Good 3 Poor 1
Here the value attached to excellent (4) is higher than the value attached to good (3), indi
cating that the response of excellent is preferred to good. However, another representation
of the ratings might be
Excellent still receives a higher value than good, but now the difference between the two
categories is 20 (100 – 80), as compared to a difference of 1 (4 – 3) when we use the
first classification. In other words, differences between categories are meaningless with
ordinal data. (We should also note that we could reverse the ordering so that, for instance,
excellent equals 40 and poor equals 100; this renumbering would not change the nature
of the data.)
EXAMPLE 1.1
In the introductory case, four questions were posed to tweens. The first question
(Q1) asked tweens to name the radio station that they listened to on the ride to the
resort, and the second question (Q2) asked tweens to rate the food quality at the
resort on a scale of 1 to 4. The tweens’ responses to these questions are shown in
Table 1.1 in the introductory case.
a. What is the scale of measurement of the radio station data?
b. How are the data based on the ratings of the food quality similar to the radio
station data? How are the data different?
c. Summarize the tweens’ responses to Q1 and Q2 in tabular form. How can the
resort use the information from these responses?
SOLUTION:
a. When asked which radio station played on the car ride to the resort, tweens re
sponded with one of the following answers: JAMN94.5, MIX 104.1, or KISS 108.
These are nominal data—the values in the data differ merely in name or label.
b. Since we can both categorize and rank the food quality data, we classify these
responses as ordinal data. Ordinal data are similar to nominal data in the sense
that we can categorize the data. The main difference between ordinal and nom
inal data is that the categories of ordinal data are ranked. A rating of 4 is better
than a rating of 3. With the radio station data, we cannot say that KISS108
is ranked higher than MIX104.1; some tweens may argue otherwise, but we
simply categorize nominal data without ranking.
c. With respect to the radio station data (Ql), we can assign 1 to JAMN94.5, 2 to
MIX 104.1, and 3 to KISS108. Counting the responses that fall into each cat
egory, we find that six tweens listened to 1, two listened to 2, and 12 listened to
3, or in tabular form:
The food quality results may be of concern to management. Just as many tweens
rated the food quality as excellent as compared to poor. Moreover, the majority
[(8 + 3)/20 = 55%] felt that the food was, at best, fair. Perhaps a more exten
sive survey that focuses solely on food quality would reveal the reason for their
apparent dissatisfaction.
As mentioned earlier, nominal and ordinal scales are used for qualitative variables. Val
ues corresponding to a qualitative variable are typically expressed in words but are coded
into numbers for purposes of data processing. When summarizing the results of a quali
tative variable, we typically count the number or calculate the percentage of persons or
objects that fall into each possible category. With a qualitative variable, we are unable to
perform meaningful arithmetic operations, such as adding and subtracting.
EXAMPLE 1.2 In the last two questions from the introductory case’s survey (Q3 and Q4), the 20 tweens
were asked: “What time should the main dining area close?” and “How much of your
own money did you spend at the lodge today?” Their responses appear in Table 1.1 in
the introductory case.
a. How are the time data classified? In what ways do the time data differ from
ordinal data? What is a potential weakness of this measurement scale?
b. What is the measurement scale of the money data? Why is it considered the
strongest form of data?
c. In what ways is the information from Q3 and Q4 useful for the resort?
SOLUTION:
a. Clock time responses, such as 3:00 pm and 3:30 pm, or 5:30 pm and 6:00 pm, are
on an interval scale. Interval data are a stronger measurement scale than ordinal
data because differences between interval-scaled values are meaningful. In this
particular example, we can say that 3:30 pm is 30 minutes later than 3:00 pm and
6:00 pm is 30 minutes later than 5:30 pm. The weakness with interval-scaled data
is that the value of zero is arbitrary. Here, with the clock time responses, we have
no apparent zero point; however, we could always arbitrarily define a zero point,
say, at 12:00 am. Thus, although differences are comparable with interval-scaled
data, ratios are meaningless due to the arbitrariness of the zero point. In other
words, it is senseless to form the ratio 6:00 pm/3:00 pm and conclude that
6:00 pm is twice as long a time period as 3:00 pm.
b. Since the tweens’ responses are in dollar amounts, this is ratio-scaled data. The
ratio scale is the strongest form of data because we can categorize and rank
values as well as calculate meaningful differences. Moreover, since there is a
natural zero point, valid ratios can also be calculated. For example, the data
show that three tweens spent $20. These tweens spent four times as much as
the three tweens that spent $5 ($20/$5 = 4).
c. A review of the clock time responses (Q3) in Table 1.1 shows that the vast
majority of the tweens would like the dining area to remain open later. In fact,
only one tween feels that the dining area should close at 3:00 pm. An inspec
tion of the money responses (Q4) in Table 1.1 indicates that only three of the
20 tweens did not spend any of his/her own money. This is very important
information. It does appear that the discretionary spending of this age group is
significant. The resort would be wise to cater to some of their preferences.
S Y N O P S I S OF I N T R O D U C T O R Y CAS E
Conceptual Review
LO 1.1 Describe the im portance of statistics.
A proper understanding of statistical ideas and concepts helps us understand more
of the real world around us, including issues in business, finance, health, and social
interactions. We must understand statistics or risk making bad decisions and costly
mistakes. A knowledge of statistics also provides the necessary tools to differentiate
between sound statistical conclusions and questionable conclusions drawn from an
insufficient number of data points, “bad” data points, incomplete data points, or just
misinformation.
LO 1.3 Explain the need for sampling and discuss various data types.
In general, we use sample data rather than population data for two main reasons:
(1) obtaining information on the entire population is expensive, and/or (2) it is impos
sible to examine every item of the population.
Cross-sectional data contain values of a characteristic of many subjects at the same
point in time or without regard to differences in time. Time series data contain values
of a characteristic of a subject over time.
2
C H A P T E R
LEARNING OBJECTIVES
After reading
this chapter
you should
Graphical Methods
LO 2.1 Summarize qualitative data by forming frequency distributions,
be able to. LO 2.3 Summarize quantitative data by form ing frequency distributions.
People often have difficulty processing information provided by data in its raw form. A useful way o f
interpreting data effectively is to condense the data w ith some kind o f visual or numerical summary.
In this chapter we present several tabular and graphical tools tha t can help you organize and pre
sent data. We first deal w ith q ualitative data by constructing freq uency distributions. We can visualize
these frequency distributions by constructing pie charts and bar charts. For quantitative data, we
again make frequency distributions. In addition to giving us an overall picture o f where the data tend
to cluster, frequency distributions using q uantitative data also show us how the data are spread out
from the lowest value to the highest value. For visual representations o f q uantitative data, we exam
ine histograms, polygons, ogives, and stem -and-leaf diagrams. Finally, we show how to construct a
scatterplot, which graphically depicts the relationship between tw o quantitative variables. We will
find that a scatterplot is a very useful tool w hen conducting correlation and regression analysis,
topics discussed in depth later in the text.
I N T R O D U C T O R Y CASE
T A B L E 2 .1 Recent Sale Price o f Houses in M ission Viejo, CA, fo r June 2010 (data in $1000s)
F R E Q U E N C Y D I S T R I B U T I O N FOR Q U A L I T A T I V E DATA
A frequency distribution for qualitative data groups data into categories and
records the number of observations that fall into each category.
To illustrate the construction of a frequency distribution with nominal data, Table 2.2
shows the weather for the month of February (2010) in Seattle, Washington.
We first note that the weather in Seattle is categorized as cloudy, rainy, or sunny. The
first column in Table 2.3 lists these categories. Initially, we use a “tally” column to
record the number of days that fall into each category. Since the first eight days of
February were rainy days, we place the first eight tally marks in the rainy category;
the ninth day of February was cloudy, so we place one tally mark in the cloudy cat
egory, and so on. Finally, we convert each category’s total tally count into its respec
tive numerical value in the frequency column. Since only one tally mark appears in
the cloudy category, we record the value 1 as its frequency. Note that if you sum the
frequency column, you obtain the sample size. A frequency distribution in its final
form does not include the tally column.
From the frequency distribution, we can now readily observe that the most common type
of day in February was rainy, since this type of day occurs with the highest frequency. In
many applications we want to compare data sets that differ in size. For example, we might
want to compare the weather in February to the weather in March. However, February has
28 days (except during a leap year) and March has 31 days. In this instance, we would
convert the frequency distribution to a relative frequency distribution. We calculate
each category’s relative frequency by dividing the respective category’s frequency by the
total number of observations. The sum of the relative frequencies should equal one, or a
value very close to one due to rounding.
Table 2.4 shows the frequency distribution in Table 2.3 converted into a relative
frequency distribution. In addition, we also show the relative frequency distribution
for the month of March. March had 3 cloudy days, 10 sunny days, and 18 rainy days.
Each of these frequencies was then divided by 31, the number of days in the month of
March.
Source: w w w .w u n d e rg ro u n d .co m .
We can easily convert relative frequencies into percentages by multiplying by 100. For
instance, the percent of cloudy days in February and March equal 3.6% and 9.7%, respec
tively. From the relative frequency distribution, we can now conclude that the weather in
Seattle in both February and March was predominantly rainy. However, the weather in
March was a bit nicer in that approximately 32% of the days were sunny, as opposed to
only 25% of the days in February.
In Adidas’ Online Annual Report 2009, net sales were reported in four regions of the EXAMPLE 2.1
world for the years 2000 and 2009 as shown in Table 2.5. Convert each region’s net
sales to its respective proportion for that year. Have the proportions of Adidas’ net
sales in each region remained the same over this 10-year period? Explain.
T A B L E 2 . 5 A d id a s' N et Sales by Region (in m illio n s o f euros, €)
SOLUTION: Over the 10-year period, Adidas’ total net sales have almost doubled.
However, it appears that the increase in net sales within each region has varied
dramatically. In order to calculate the proportions of Adidas’ net sales for each region,
we take each region’s net sales and divide by the year’s total sales, as shown in Table 2.6.
TA B LE 2 .6 P ro p o rtio n o f A d id a s' N et Sales by Region
Total = 1 Total = 1
Once we convert the data to proportions, we see significant changes in the propor
tion of net sales allocated to each region. In 2009, Europe still has the highest per
centage of net sales at 42.3%; however, this percentage has fallen over the 10-year
period. A large decline took place in the percentage of net sales in North America,
from 32.8% to 22.8%, compared to significant increases in the percentages of net
sales in Asia, from 15.1% to 25.2%, and in Latin America, from 2.9% to 9.7%.
In short, there has been considerable movement in the percentage of Adidas’ net
sales allocated to each region over the 10-year period. This type of information
can help Adidas when making important marketing decisions.
In order to construct a pie chart, first draw a circle. Then cut the circle into slices, or sectors,
such that each sector is proportional to the size of the category you wish to display. For
instance, Table 2.6 shows that Europe accounted for 49.2% of Adidas’ net sales in 2000.
Since a circle contains 360 degrees, the portion of the circle representing Europe encom
passes 0.492 × 360 = 177.1 degrees; thus, almost half of the circle should reflect Europe’s
contribution to sales. Similar calculations for the other three regions in 2000 yield:
North America: 0.328 × 360 = 118.1 degrees
Asia: 0.151 × 360 = 54.4 degrees
Latin America: 0.029 × 360 = 10.4 degrees
The same methodology can be used to calculate each region’s contribution to net sales for
the year 2009. Figure 2.1 shows the resulting pie charts.
FIG U R E 2 . 1
Pie c h a rts for A didas'
n e t sales.
Using Excel to Construct a Pie Chart
Excel offers various options for displaying a pie chart. To replicate the pie chart in
Figure 2.1a, follow these steps:
A. Open the Adidas Sales data (Table 2.5) from the text website into an Excel
spreadsheet.
B. Select the categorical names and respective frequencies from the year 2000. Leave
out the heading (top row); see Figure 2.2 below.
C. From the menu choose Insert > Pie > 2-D Pie. From the options given, choose the
graph on the top left.
D. In order to give the pie chart category names and their respective percentages, from
the menu choose Layout > Data Labels > More Data Label Options. Under
Label Options, deselect “Value” and select “Category Name” and “Percentage.”
Another way to graphically depict qualitative data is to construct a bar chart. When con
structing a bar chart, first place each category on the horizontal axis and then mark the
vertical axis with an appropriate range of values for either frequency or relative frequency.
The height of each bar is equal to the frequency or the relative frequency of the correspond
ing category. Typically, you should leave space between categories to improve clarity.
G R A P H I C A L D I S P L A Y OF Q U A L I T A T I V E D A T A: BAR C H A R T S
A bar chart depicts the frequency or the relative frequency for each category of the
qualitative data as a bar rising vertically from the horizontal axis.
Figure 2.3 shows a relative frequency bar chart for the Adidas net sales example. It is
particularly useful because we can group net sales by region, emphasizing the rise in the
proportion of sales in Asia and Latin America versus the fall in the proportion of sales in
Europe and North America over the 10-year period.
B. Select the categorical names and respective relative frequencies for the years 2000
and 2009. Leave out the heading (top row); see Figure 2.4.
C. Choose Insert > Column > 2-D Column. From the options given, choose the
graph on the top left. (This will create a vertical bar chart. If you want to construct a
horizontal bar chart, choose Insert > Bar > 2-D Bar.)
D. In the legend to the right of the bar chart, Excel labels the data for the year 2000 as
“Series 1” and the data for the year 2009 as “Series 2” by default. In order to edit
the legend, select the legend and choose Design > Select Data. From the Legend
Entries, select “Series 1,” then select Edit, and under Series Name, type the new
name of 2000. Follow the same steps to rename “Series 2” to 2009.
Source: U.S. E n e rg y In fo rm a tio n A d m in is tra tio n . Source: h ttp ://w w w .fin a n ce .ya h o o .co m .
• The scale on the vertical axis should begin at zero. Moreover, the vertical axis
should not be given a very high value as an upper limit. In these instances, the data
may appear compressed so that an increase (or decrease) of the data is not as ap
parent as it perhaps should be. Conversely, the axis should not be stretched so that
an increase (or decrease) of the data appears more pronounced than warranted. For
example, Figure 2.5(a) plots the daily price for a barrel of crude oil for the first
quarter of 2011 (the data labeled Crude Oil are available on the text website). Due
to Middle East unrest, the price of crude oil rose from a low of $83.13 per barrel
to a high of $ 106.19 per barrel, or approximately However,
since Figure 2.5(a) uses a high value as an upper limit on the vertical axis ($325),
the rise in price appears dampened. Figure 2.5(b) charts the daily closing stock price
for Johnson & Johnson (JNJ) for the week of April 4, 2011 (the data labeled J N J are
available on the text website). It is true that the stock price declined over the week
from a high of $60.15 to a low of $59.46; this amounts to a $0.69 decrease or an ap
proximate 1% decline. However, since the vertical axis is stretched, the drop in stock
price appears more dramatic.
• When creating a bar chart, each bar should be of the same width. Increasing (or de
creasing) bar widths creates distortions. The same principle holds in the next section
when we discuss histograms.
EXERCISES 2. 1
1. A local restaurant is com m itted to providing its patrons w ith chooses Excellent, Good, Fair, o r Poor. Over th e past m o n th ,
the best dining experience possible. On a recent survey, the th e responses to th is qu e stion fro m firs t-tim e p a tie nts w ere:
restaurant asked patrons to rate the quality of their entrees.
Fair Good Fair Excellent
The responses ranged from 1 to 5, where 1 indicated a
Good Good Good Poor
disappointing entree and 5 indicated an exceptional entree.
The results o f the survey are as follows: Excellent Excellent Poor Good
a. Construct a frequency and a relative frequency a. Construct a frequency and a relative frequency
distribution that summarizes the survey's results. distribution that summarizes the responses to the
b. Are patrons generally satisfied w ith the quality o f their questionnaire.
entrees? Explain. b. What is the most common response to the
2. First-time patients at North Shore Family Practice are questionnaire? How would you characterize the health
required to fill out a questionnaire that gives the doctor an o f first-tim e patients at this medical practice?
overall idea of each patient's health. The first question is: 3. A survey asked chief executives at leading U.S. firms the
"In general, w hat is the quality of your health?" The patient follow ing question: "Where do you expect the U.S. economy
to be 12 months from now?" A representative sample o f their Response Relative Frequency
responses appears below:
Good jobs 0.37
Same Same Same Better Worse Affordable homes 0.15
Same Same Better Same Worse Top schools 0.11
Same Better Same Better Same Low crime 0.23
Worse Same Same Same Worse Things to do 0.14
Same Same Same Better Same
Source: CNNM oney.com , July 13,2010.
a. Construct a frequency and a relative frequency a. Construct the corresponding frequency distribution.
distribution tha t summarizes the responses to the How many of the respondents chose "low crime" as
survey. Where did most chief executives expect the U.S. the most important criteria when choosing where
economy to be in 12 months? to live?
b. Construct a pie chart and a bar chart to summarize your b. Construct a bar chart for these data.
results. 7. What is the perfect summer trip? A National Geographic
4. AccuWeather.com reported the follow ing weather delays at Kids survey (AAA Horizons, April 2007) asked this question
these major U.S. airline hubs for July 2 1 , 2010: to 316 children ages 8 to 14. Their responses are given in the
City Delay City Delay following frequency distribution.
Atlanta PM Delays Mpls./St. Paul None Top Vacation Choice Frequency
Chicago None New York All Day Delays Cruises 140
Dallas/Ft. Worth None Orlando None Beaches 68
Denver All Day Delays Philadelphia All Day Delays Amusement Parks 68
Detroit AM Delays Phoenix None Big Cities 20
Houston All Day Delays San Francisco AM Delays Lakes 12
Las Vegas All Day Delays Salt Lake City None Summer Camp 8
Los Angeles AM Delays Seattle None a. Construct a relative frequency distribution. What
Miami AM Delays Washington All Day Delays percentage of the responses cited "Cruises" as the perfect
summer trip?
a. Construct a frequency and a relative frequency
distribution that summarizes the delays at major U.S. b. Construct a bar chart for these data.
hubs. W hat was th e m ost co m m o n ty p e o f delay? Explain. 8. The following table lists U.S. revenue (in $ billions) of the
b. Construct a pie chart and a bar chart to summarize your major car-rental companies.
results. Car-Rental Company Revenue in 2009
5. Fifty pro-football rookies were rated on a scale o f 1 to 5, Enterprise $10.7
based on performance at a training camp as well as on Hertz 4.7
past performance. A ranking o f 1 indicated a poor prospect Avis Budget 4.0
whereas a ranking o f 5 indicated an excellent prospect. The
Dollar Thrifty 1.5
follow ing frequency distribution was constructed.
Other 1.0
Rating Frequency
Source: The W all Street Journal, July 30,2010.
1 4
a. Construct a relative frequency distribution.
2 10
b. Hertz accounted for what percentage of sales?
3 14
c. Construct a pie chart for these data.
4 18
9. A survey conducted by CBS News asked 829 respondents
5 4
which of the following events will happen first. The
a. H ow m any o f th e rookies received a ratin g o f 4 o r better? responses are summarized in the following table:
H ow m any o f th e rookies received a ra tin g o f 2 o r worse?
Cure for cancer found 40%
b. Construct the corresponding relative frequency
End of dependence on oil 27%
distribution. W hat percent received a rating o f 5?
Signs of life in outer space 12%
c. Construct a bar chart for these data.
Peace in Middle East 8%
6. A recent survey asked 5,324 individuals: "What's most
Other 6%
im portant to you when choosing where to live?" The
None will happen 7%
responses are shown in the follow ing relative frequency
distribution. Source: Vanity Fair, D ecem ber 2009.
a. Construct a bar chart and a pie chart for these data. period, almost tripling in price. Does the figure reflect this
b. How many people think that a cure for cancer will be growth? If not, why not?
found first?
10. A 2010 poll conducted by NBC asked respondents who
would win Super Bowl XLV in 2011. The responses by 20,825
people are summarized in the following table.
Team Number of Votes
Atlanta Falcons 4,040
New Orleans Saints 1,880
Houston Texans 1,791
Dallas Cowboys 1,631
Minnesota Vikings 1,438 12. Annual sales at a small pharmaceutical firm have been rather
Indianapolis Colts 1,149 stagnant over the most recent five-year period, exhibiting
only 1.2% growth over this tim e frame. A research analyst
Pittsburgh Steelers 1,141
prepares the accompanying graph for inclusion in a sales
New England Patriots 1,095
report.
Green Bay Packers 1,076
Others
SOLUTION: The frequency distribution shows that house prices ranged from
$300,000 up to $800,000 over this time period. The most houses (14) sold in the
$500,000 up to $600,000 range. Note that only four houses sold in the lowest price
range and only two houses sold at the highest price range.
It turns out that reading and understanding a frequency distribution is actually easier than
forming one. When we constructed a frequency distribution with qualitative data, the raw
data could be categorized in a well-defined way. With quantitative data, we must make
certain decisions about the number of classes, as well as the width of each class. We do
not apply concrete rules when we define the classes in Table 2.7; however, we are able to
follow several guidelines.
Generally, the width of each class is the same for each class interval. If the class
width varied, comparisons between the numbers of observations in different inter
vals would be misleading. Moreover, it is preferable to define class limits that are
easy to recognize and interpret.
Suppose we conclude, as we do in Table 2.7, that we should have five classes in
the frequency distribution for the house-price data. Applying the class-width for
mula with the largest value of 735 and the smallest value of 330 (from Table 2.1)
yields Table 2.10 shows the frequency distribution with five classes
and a class width of 81.
Again, this is a valid frequency distribution, but it proves unwieldy. Recall that one
major goal in forming a frequency distribution is to provide more clarity in inter
preting the data. Grouping the data in this manner actually makes analyzing the
data more difficult. In order to facilitate interpretation of the frequency distribution,
it is best to define class limits with ease of recognition in mind. To this end, and as
initially shown in Table 2.7, we set the lower limit of the first class at 300 (rather
than 330) and obtain the remaining class limits by successively adding 100 (rather
than 81).
Once we have clearly defined the classes for a particular data set, the next step
is to count and record the number of data points that fall into each class. As we did
with the construction of a qualitative frequency distribution, we usually include a
tally column to aid in counting (see Table 2.11), but then we remove this column in
the final presentation of the frequency distribution. For instance, in Table 2.1, the
first data point, 430, falls in the second class, so we place a tally mark in the sec
ond class; the next value of 520 falls in the third class, so we place a tally mark in
the third class, and so on. The frequency column shows the numerical value of the
respective tally count. Since four tally marks appear in the first class, we record the
value 4 as its frequency—the number of observations that fall into the first class.
One way to ensure that we have included all the data points in the frequency distri
bution is to sum the frequency column. This sum should always equal the population
or sample size.
TABLE 2 .1 1 Constructing Frequency Distributions for the House-Price Data
Class (in $1000s) Tally Frequency Cumulative Frequency
300 up to 400 4 4
400 up to 500 11 4 + 11 = 15
500 up to 600 14 4 + 1 1 + 14 = 29
600 up to 700 5 4 + 11 + 14 + 5 = 34
700 up to 800 2 4+11 +14 + 5 + 1 =36
Total = 36
A frequency distribution indicates how many observations (in this case house prices) fall
within some range. However, we might want to know how many observations fall below
the upper limit of a particular class. In these cases, our needs are better served with a
cumulative frequency distribution.
The last column of Table 2.11 shows values for cumulative frequency. The cumula
tive frequency of the first class is the same as the frequency of the first class, that is, the
value 4. However, the interpretation is different. With respect to the frequency column,
the value 4 tells us that four of the houses sold in the $300,000 up to $400,000 range.
For the cumulative frequency column, the value 4 tells us that four of the houses sold
for less than $400,000. To obtain the cumulative frequency for the second class, we add
its frequency, 11, with the preceding frequency, 4, and obtain 15. This tells us that 15
of the houses sold for less than $500,000. We solve for the cumulative frequencies of
the remaining classes in a like manner. Note that the cumulative frequency of the last
class is equal to the sample size of 36. This indicates that all 36 houses sold for less than
$800,000.
EXAMPLE 2.3
Using Table 2.11, how many of the houses sold in the $500,000 up to $600,000
range? How many of the houses sold for less than $600,000?
SOLUTION: From the frequency distribution, we find that 14 houses sold in the
$500,000 up to $600,000 range. In order to find the number of houses that sold
for less than $600,000, we use the cumulative frequency distribution. We readily
observe that 29 of the houses sold for less than $600,000.
Suppose we want to compare house prices in Mission Viejo, California, to house prices
in another region of the United States. Just as for qualitative data, when making com
parisons between two quantitative data sets—especially if the data sets are of different
sizes—a relative frequency distribution tends to provide more meaningful information as
compared to a frequency distribution.
The second column of Table 2.12 shows the construction of a relative frequency dis
tribution from the frequency distribution in Table 2.11. We take each class’s frequency
and divide by the total number of observations. For instance, we observed four houses
that sold in the lowest range of $300,000 up to $400,000. We take the class frequency of
4 and divide by the sample size, 36, and obtain 0.11. Equivalently, we can say 11 % of the
houses sold in this price range. We make similar calculations for each class and note that
when we sum the column of relative frequencies, we should get a value of one (or, due to
rounding, a number very close to one).
Relative
Class (in $1000s) Frequency Cumulative Relative Frequency
300 up to 400 4/36 = 0.11 0.11
400 up to 500 11/36 = 0.31 0.11 +0.31 =0.42
500 up to 600 14/36 = 0.39 0.11 + 0.31 + 0.39 = 0.81
600 up to 700 5/36 = 0.14 0.11 +0.31 +0.39 + 0.14 = 0.95
700 up to 800 2/36 = 0.06 0.11 +0.31 + 0.39 + 0.17 + 0 .0 6 = 1
Total = 1
(approximately)
The last column of Table 2.12 shows cumulative relative frequency. The cumulative
relative frequency for a particular class indicates the proportion or fraction of the ob
servations that fall below the upper limit of that particular class. We can calculate the
cumulative relative frequency of each class in one of two ways: (1) we can sum succes
sive relative frequencies, or (2) we can divide each class’s cumulative frequency by the
sample size. In Table 2.12 we show the first way. The value for the first class is the same
as the value for its relative frequency, that is, 0.11. For the second class we add 0.31 to
0.11 and obtain 0.42; this value indicates that 42% of the house prices were less than
$500,000. We continue calculating cumulative relative frequencies in this manner until
we reach the last class. Here, we get the value one, which means that 100% of the houses
sold for less than $800,000.
RELATIVE A N D C U M U L A T IV E RELATIVE
FREQUENCY DISTRIBUTIONS
For quantitative data, a relative frequency distribution identifies the proportion
(or the fraction) of values that fall into each class, that is,
EXAMPLE 2 .4
Using Table 2.12, what percent of the houses sold for at least $500,000 but not
more than $600,000? What percent of the houses sold for less than $600,000? What
percent of the houses sold for $600,000 or more?
SOLUTION: The relative frequency distribution indicates that 39% of the houses
sold for at least $500,000 but not more than $600,000. Further, the cumulative
relative frequency distribution indicates that 81% of the houses sold for less than
$600,000. This result implies that 19% sold for $600,000 or more.
Visualizing Frequency Distributions for Quantitative Data
LO 2 .4 Histograms and polygons are graphical depictions of frequency and relative frequency
Construct and distributions. The advantage of a visual display is that we can quickly see where most of
interpret histograms, the observations tend to cluster, as well as the spread and shape of the data. For instance,
polygons, and ogives. histograms and polygons may reveal whether or not the distribution is symmetrically
shaped.
G R A P H I C A L D IS P L A Y OF Q U A N T I T A T I V E DATA: H I S T O G R A M S
A histogram is a series of rectangles where the width and height of each rectangle
represent the class width and frequency (or relative frequency) of the respective class.
For quantitative data, a histogram is essentially the counterpart to the bar chart we use for
qualitative data. When constructing a histogram, we mark off the class limits along the
horizontal axis. The height of each bar represents either the frequency or relative frequency
for each class. No gaps appear between the interval limits. Figure 2.6 shows a histogram
for the frequency distribution of house prices shown in Table 2.7. A casual inspection of
the histogram reveals that the selling price of houses in this sample ranged from $300,000
to $800,000; however, most house prices fell in the $500,000 to $600,000 range.
The only difference between a frequency histogram and a relative frequency histogram
is the unit of measurement on the vertical axis. For the frequency histogram, we use the
frequency of each class to represent the height; for the relative frequency histogram we use
the proportion (or the fraction) of each class to represent the height. In a relative frequency
histogram, the area of any rectangle is proportional to the relative frequency of observations
falling into that class. Figure 2.7 shows the relative frequency histogram for house prices.
The histogram in Figure 2.8(a) shows a symmetric distribution. If the edges were
smoothed, this histogram would approximate the bell-shaped normal distribution. In
Figure 2.8(b), the histogram shows a positively skewed, or skewed to the right, distri
bution with a long tail extending to the right. This attribute reflects the presence of a
small number of relatively large values. Finally, the histogram in Figure 2.8(c) indicates
a negatively skewed, or skewed to the left, distribution since it has a long tail extending
off to the left. Data that follow a negatively skewed distribution have a small number of
relatively small values.
Though not nearly as skewed as the data exhibited in Figure 2.8(b), the house-price
data in Figure 2.7 exhibit slight positive skew. This is the result of a few, relatively ex
pensive homes in the city. It is common for distributions of house prices and incomes to
exhibit positive skewness.
C. From the menu choose Data > Data Analysis > Histogram > OK. (Note: If you
do not see the Data Analysis option under Data, you must add-in this option. Go to
the Office Button and choose Excel Options > Add-Ins. Then choose the box to
the left of Analysis Toolpak, choose Go, and then click OK. If you have installed
this option properly, you should now see Data Analysis under Data.)
D. In the Histogram dialog box (see Figure 2.10), under Input Range, select the data.
Excel uses the term “bins” for the class limits. If we leave the Bin Range box empty,
Excel creates evenly distributed intervals using the minimum and maximum values
of the input range as end points. This methodology is rarely satisfactory. In order to
construct a histogram that is more informative, we use the upper limit of each class
as the bin values. Under Bin Range, we select the Class Limits data. (Check the
Labels box if you have included the names House Price and Class Limits as part of
the selection.) Under Output Options we choose Chart Output, then click OK.
E. Since Excel leaves spaces between the rectangles, we right-click on any of the rect
angles, choose Format Data Series and change the Gap Width to 0, then choose
Close. In the event that the given class limits do not include all the data points,
Excel automatically adds another interval labeled “More” to the resulting frequency
distribution and histogram. Since we observe zero observations in this interval for
this example, we delete this interval for expositional purposes. Excel also defines its
classes by excluding the value of the lower limit and including the value of the upper
class limit for each interval. For example, if the value 400 appeared in the house-
price data, Excel would have accounted for this observation in the first class. If any
upper-limit value appeared in the house-price data, we would have adjusted the class
limits in the Bin Range to 399, 499, etc., so that Excel’s frequency distribution and
histogram would be consistent with those that we constructed in Table 2.11 and Fig
ure 2.6. Further formatting regarding colors, axes, grids, etc. can be done by select
ing Layout from the menu.
G R A P H I C A L D I S P L A Y OF Q U A N T I T A T I V E D A T A : P O L Y G O N S
A polygon connects a series of neighboring points where each point represents the
midpoint of a particular class and its associated frequency or relative frequency.
If we choose to construct a polygon for the house-price data, we first calculate the mid
point of each interval; thus, the midpoint for the first interval is an d
similarly, the midpoints for the remaining intervals are 450, 550, 650, and 750. We treat
each midpoint as the x-coordinate and the respective frequency (or relative frequency)
as the y-coordinate. After plotting the points, we connect neighboring points. In order to
close off the graph at each end, we add one interval below the lowest interval (so, 200 up
to 300 with midpoint 250) and one interval above the highest interval (so, 800 up to 900
with midpoint 850) and assign each of these classes zero frequencies. Table 2.13 shows
the relevant coordinates for plotting a polygon using the house-price data. We chose to
use relative frequency to represent the v-coordinate.
Figure 2.12 plots a relative frequency polygon for the house-price data. Here the distribu
tion appears to approximate the bell-shaped distribution discussed earlier. Only a careful
inspection of the right tail suggests that the data are slightly positively skewed.
G R A P H I C A L DISPLAY OF Q U A N T I T A T I V E DATA: O G IV E S
An ogive is a graph that plots the cumulative frequency or the cumulative relative
frequency of each class against the upper limit of the corresponding class.
An ogive differs from a polygon in that we use the upper limit of each class as the x-
coordinate and the cumulative frequency or cumulative relative frequency of the cor
responding class as the y-coordinate. After plotting the points, we connect neighboring
points. Lastly, we close the ogive only at the lower end by intersecting the x-axis at the
lower limit of the first class. Table 2.14 shows the relevant coordinates for plotting an
ogive using the house-price data. We choose to use cumulative relative frequency as the
y-coordinate. The use of cumulative frequency would not change the shape of the ogive,
just the unit of measurement on the y-axis.
x-coordinate y-coordinate
Classes (upper limit) (cumulative relative frequency)
(Lower end) 300 0
300-400 400 0.11
Figure 2.14 plots the ogive for the house-price data. In general, we can use an ogive to
approximate the proportion of values that are less than a specified value on the horizontal
axis. Consider an application to the house-price data in Example 2.5.
FIGURE 2 .1 4 Ogive for the house-price data
EXAMPLE 2.5
Using Figure 2.14, approximate the percentage of houses that sold for less than
$550,000.
SOLUTION: Draw a vertical line that starts at 550 and intersects the ogive. Then
follow the line to the vertical axis and read the value. You can conclude that
approximately 60% of the houses sold for less than $550,000.
EXERCISES 2.2
4 10 8 7 6 10 11 14 13 14
3 9 8 5 7 6 10 3 11 11 a. Construct a relative frequency distribution. Graph a
8 8 4
relative frequency histogram.
5 5 12 12 3 8 8
10 -9 28 14
b. Construct a cumulative frequency distribution and a
-5 9 11 5 8 -3
cumulative relative frequency distribution.
33 -4 2 3 22 25 5 29 26 0
c. What percent of the observations are at least 30 but
-8 -5 0 15 -4 35 21 15 19 23
less than 40? What percent of the observations are less
4 6 -2 12 24 36 15 3 -5 2 than 40?
16. Consider the follow ing frequency distribution: Applications
Class Frequency 20. Kiplinger's (August 2007) lists the assets (in billions o f $) for
1000 up to 1100 2 the 20 largest stock mutual funds (ranked by size) as follows:
1100 up to 1200 7 $99.8 49.7 86.3 109.2 56.9
1200 up to 1300 3 88.2 44.1 58.8 176.7 49.9
1300 up to 1400 4 61.4 128.8 53.6 95.2 92.5
55.0 96.5 45.3 73.0 70.9
a. Construct a relative frequency distribution. What
percent o f the observations are at least 1100 but less a. Construct a frequency distribution using classes o f 40 up
than 1200? to 7 0 , 70 up to 100, etc.
b. Construct a cumulative frequency distribution and a b. Construct the relative frequency distribution, the
cumulative relative frequency distribution. How many o f cumulative frequency distribution, and the relative
the observations are less than 1300? cumulative frequency distribution.
c. Graph a frequency histogram. c. How many o f the funds had assets o f at least $100 but
17. Consider the follow ing cum ulative frequency distribution: less than $130 (in billions)? How many o f the funds had
assets less than $160 (in billions)?
Class Cumulative Frequency
d. W hat percent o f the funds had assets o f at least $70 but
15 up to 25 30
less than $100 (in billions)? W hat percent o f the funds
25 up to 35 50
had assets less than $130 (in billions)?
35 up to 45 120
e. Construct a histogram. Comment on the shape o f the
45 up to 55 130 distribution.
a. Construct a frequency distribution. How many 21. The numbers o f texts sent by 25 13-year-olds over the past
observations are at least 35 but less than 45? m onth were as follows:
b. Graph a frequency histogram. 630 516 892 643 627 510 937 909 654
c. What percent o f the observations are less than 45? 817 760 715 605 975 888 912 952 701
18. Consider the follow ing relative frequency distribution: 744 793 852 504 562 670 685
Class Relative Frequency a. Construct a frequency distribution using classes of 500
-2 0 up to —10 0.04 up to 600,600 up to 700, etc.
-1 0 u p to 0 0.28 b. Construct the relative frequency distribution, the
0 up to 10 0.26 cumulative frequency distribution and the relative
10 up to 20 0.22 cumulative frequency distribution.
20 up to 30 0.20 c. How many of the 13-year-olds sent at least 600 but less
than 700 texts? How many sent less than 800 texts?
a. Suppose this relative frequency distribution is based
d. What percent of the 13-year-olds sent at least 500 but
on a sample o f 50 observations. Construct a frequency
less than 600 texts? What percent of the 13-year-olds
distribution. How many o f the observations are at least
sent less than 700 texts?
—10 but less thanO?
e. Construct a polygon. Comment on the shape of the
b. Construct a cumulative frequency distribution. How
distribution.
many o f the observations are less than 20?
22. AccuWeather.com listed the following high temperatures
c. Graph a relative frequency polygon.
(in degrees Fahrenheit) for 33 European cities on
19. Consider the follow ing cumulative relative frequency July 21, 2010.
distribution.
75 92 81 85 90 73 94 95 81 64 85
Class Cumulative Relative Frequency
62 84 85 81 86 90 79 74 90 91 95
150 up to 200 0.10
88 87 81 73 76 86 90 83 75 92 83
200 up to 250 0.35
a. Construct a frequency distribution using classes o f 60 up
250 up to 300 0.70
to 7 0 , 70 up to 80, etc.
300 up to 350 1
b. Construct the relative frequency, the cumulative
a. Construct a relative frequency distribution. What frequency, and the relative cumulative frequency
percent of the observations are at least 250 but less distributions.
than 300? c. How many o f the cities had high temperatures less
b. Graph an ogive. than 80°?
d. What percent of the cities had high temperatures of at c. Construct a histogram. Comment on the shape of the
least 80° but less than 90°? What percent of the cities had distribution.
high temperatures less 90°? 25. The following relative frequency distribution summarizes the
e. Construct a polygon. Comment on the shape of the ages of women who had a child in the last year.
distribution.
Ages Relative Frequency
23. The following table lists a portion of the average math SAT
scores for each state for the year 2009.The complete data set can 15 up to 20 0.10
be found on the text website and is labeled Math SAT2009. 20 up to 25 0.25
25 up to 30 0.28
State SAT
30 up to 35 0.24
Alabama 552
35 up to 40 0.11
Alaska 516
40 up to 45 0.02
Source: The Statistical A bstract o f the United States, 2010.
Wyoming 568
Source: w w w .co lle g eb oa rd .com . a. Assume the relative frequency distribution is based on a
a. Construct a frequency distribution using classes of 450 sampling of 2,000 women. Construct the corresponding
to 500,501 to 550, etc. How many of the states had frequency distribution, cumulative frequency distribution,
scores between 551 and 600? and cumulative relative frequency distribution.
b. Construct the relative frequency, the cumulative b. What percent of the women were at least 25 but less
frequency, and the relative cumulative frequency than 30 years old? What percent of the women were
distributions. younger than 35 years old?
c. How many of the states had math SAT scores of 550 or c. Construct a relative frequency polygon. Comment on the
less? shape of the distribution.
d. What percent of the states had math SAT scores between d. Construct an ogive. Using the graph, approximate the
551 and 600? What percent of the states had mean SAT age of the middle 50% of the distribution.
scores of 550 or less? 26. The manager of a nightclub near a local university recorded
e. Construct a histogram. Comment on the shape of the the ages of the last 100 guests in the following cumulative
distribution. frequency distribution.
24. Fifty cities provided information on vacancy rates (in percent) Ages Cumulative Frequency
in local apartments in the following frequency distribution. 18 up to 22 45
Vacancy Rate (in percent) Frequency 22 up to 26 70
0 up to 3 5 26 up to 30 85
3 up to 6 10 30 up to 34 96
6 up to 9 20 34 up to 38 100
9 up to 12 10
a. Construct the corresponding frequency, relative frequency,
12 up to 15 5
and cumulative relative frequency distributions.
a. Construct the corresponding relative frequency b. How many of the guests were at least 26 but less than
distribution, cumulative frequency distribution, and 30 years old? What percent of the guests were at least
cumulative relative frequency distribution. 22 but less than 26 years old? What percent of the guests
b. How many of the cities had a vacancy rate less than were younger than 34 years old? What percent were
12%? What percent of the cities had a vacancy rate of at 34 years or older?
least 6% but less than 9%? What percent of the cities had c. Construct a histogram. Comment on the shape of the
a vacancy rate of less than 9%? distribution.
2 .3 S te m - a n d -L eaf D ia g ra m s
John Tukey (1915–2000), a well-known statistician, provided another visual method for LO 2 .5
displaying quantitative data. A stem-and-leaf diagram is often a preliminary step when Construct and
analyzing a data set. It is useful in that it gives an overall picture of where the data are interpret a stem -
centered and how the data are dispersed from the center. and-leaf diagram.
G R A P H I C A L D I S P L A Y OF Q U A N T I T A T I V E D A T A:
STEM -AND-LEAF DIAGRAMS
A stem-and-leaf diagram is constructed by separating each value of a data set into
two parts: a stem, which consists of the leftmost digits, and a leaf, which consists
of the last digit.
EXAMPLE 2 .6
Table 2.15 shows the ages of the 25 wealthiest people in the world; these data labeled
Wealthiest People are also available on the text website. Construct and interpret a
stem-and-leaf diagram.
TABLE 2.15 Wealthiest People in the World, 2010
Name Age Name Age
Carlos Slim Helu 70 Li Ka-shing 81
William Gates III 54 Jim Walton 62
Warren Buffet 79 Alice Walton 60
Mukesh Ambani 52 Liliane Bettencourt 87
Lakshmi Mittal 59 S. Robson Walton 66
Lawrence Ellison 65 Prince Alwaleed Alsaud 54
Bernard Arnault 61 David Thomson 52
Eike Batista 53 Michael Otto 66
Amancio Ortega 74 Lee Shau Kee 82
Karl Albrecht 90 Michael Bloomberg 68
IngvarKamprad 83 Sergey Brin 36
Christy Walton 55 Charles Koch 74
Stefan Persson 62
Source: w w w .fo rb es.co m /lists/20 1 0 .
SOLUTION: For each age we first decide that the number in the tens spot will denote
the stem, thus leaving the number in the ones spot as the leaf. We then identify the
lowest and highest values in the data set. Sergey Brin is the youngest member of
this group at 36 years of age (stem: 3, leaf: 6) and Karl Albrecht is the oldest at
90 years of age (stem: 9, leaf: 0). These values give us the first and last values in
the stem. This means our stems will be 3, 4, 5, 6, 7, 8, and 9, as shown in Panel A
of Table 2.16.
TABLE 2.16 Constructing a Stem-and-Leaf Diagram for Example 2.6
Panel A Panel B Panel C
Stem Leaf Stem Leaf Stem Leaf
3 3 6 3 6
4 4 4
5 5 4293542 5 2234459
6 6 51220668 6 01225668
7 0 7 0944 7 0449
8 8 3172 8 12 3 7
9 9 0 9 0
We then begin with the wealthiest man in the world, Carlos Slim Helu, whose age
of 70 gives us a stem of 7 and a leaf of 0. We place a 0 in the row corresponding
to a stem of 7, as shown in Panel A of the table. We continue this process with all
the other ages and obtain the values in Panel B. Finally, in Panel C we arrange each
individual leaf row in ascending order. Panel C is the actual stem-and-leaf display.
The stem-and-leaf diagram (Panel C) presents the original 25 values in a more
organized form. From the diagram we can readily observe that the ages range from
36 to 90. Wealthy individuals in their sixties make up the greatest group in the sam
ple with eight members, while those in their fifties place a close second, accounting
for seven members. We also note that the distribution is not perfectly symmetric.
A stem-and-leaf diagram is similar to a histogram turned on its side with the added
benefit of retaining the original values.
EXERCISES 2.3
2.4 Scatterplots
All of the tabular and graphical tools presented thus far have focused on describ LO 2.6
ing one variable. However, in many instances we are interested in the relationship Construct and
between two variables. People in virtually every quantitative discipline examine interpret a scatterplot.
how one variable may systematically influence another variable. Consider, for in
stance, how
• Incomes vary with education.
• Sales vary with advertising expenditures.
• Stock prices vary with corporate profits.
• Crop yields vary with the use of fertilizer.
• Cholesterol levels vary with dietary intake.
• Weight varies with exercise.
SCATTERPLOTS
A scatterplot is a graphical tool that helps in determining whether or not two vari
ables are related in some systematic way. Each point in the diagram represents a pair
of known or observed values of the two variables.
When constructing a scatterplot, we generally refer to one of the variables as x and rep
resent it on the horizontal axis and the other variable as y and represent it on the vertical
axis. We then plot each pairing: (x 1, y 1), (x2, y’2), etc. Once the data are plotted, the graph
may reveal that
• A linear relationship exists between the two variables;
• A curvilinear relationship exists between the two variables; or
• No relationship exists between the two variables.
For example, Figure 2.16(a) shows points on a scatterplot clustered together in a straight,
upward-sloping line; we infer that the two variables have a positive linear relationship.
Part (b) depicts a positive curvilinear relationship; as x increases, y tends to increase at an
increasing rate. The points in part (c) are scattered with no apparent pattern; thus, there is
no relationship between the two variables.
EXAMPLE 2.7
A social scientist wants to analyze the relationship between educational attainment
and salary. He collects the following data, where “education” refers to years of higher
education and “income” is the individual’s annual salary in thousands of dollars.
Construct and interpret a scatterplot.
Individual Education Income
1 3 45
2 4 56
3 6 85
4 2 35
5 5 55
6 4 48
7 8 100
8 0 38
SOLUTION: We let x and y denote education and income, respectively. We plot the
first individual’s pairing as (3, 45), the second individual’s pairing as (4, 56), and so
on. The graph should resemble Figure 2.17.
As expected, we observe a positive relationship between the two variables; that is,
when education increases, income tends to increase.
FIGURE 2.18
Constructing a scatterplot
with Excel
EXERCISES 2.4
The tabular and graphical tools introduced in this chapter are the starting point for most
studies and reports that involve statistics. They can help you organize data so you can see
patterns and trends in the data, which can then be analyzed by the methods described in
later chapters of this book. In this section, we present an example of using tabular and
graphical methods in a sample report. Each of the remaining chapters contains a sample
report incorporating the concepts developed in that respective chapter.
Camilla Walford is a newly hired journalist for a national
newspaper. One of her first tasks is to analyze gas prices in the
United States during the week of the Fourth of July holiday.
She collects average gas prices for the 48 contiguous states and
the District of Columbia (DC), a portion of which is shown in
Table 2.17. The complete data set can be found on the text
website and is labeled Gas Prices, Summer 2010.
TABLE 2 .1 7 U.S. Gas Prices, July 2, 2010
Wyoming 2.77
Historically, in the United States, many people choose to take some time off during the Sample
Fourth of July holiday period and travel to the beach, the lake, or the mountains. The
roads tend to be heavily traveled, making the cost of gas a concern. The following report Report— Gas
provides an analysis of gas prices across the nation over this holiday period.
The analysis focuses on the average gas price for the 48 contiguous states and the Prices across
District of Columbia (henceforth, referenced as 49 states for ease of exposition). The
range of gas prices is from a low of $2.52 per gallon (South Carolina) to a high of the United
$3.15 per gallon (California). To find out how gas prices are distributed between these
extremes, the data have been organized into several frequency distributions as shown in States
Table 2.A. For instance, most states (17 of the 49) have an average gas price from $2.70
up to $2.80 per gallon. Equivalently, looking at the relative frequency column, 35% of the
states have an average price in this range. The cumulative frequency column indicates that
35 states (out of a total of 49) have an average price less than $2.80 per gallon. Finally,
the last column shows that the average price in 72% of the states (approximately three-
quarters of the sample) is less than $2.80 per gallon.
TABLE 2 .A Frequency Distributions for Gas Prices in the United States, July 2,2010
Figure 2.A shows a histogram for gas prices, which graphs the frequency distribution
from Table 2. A. This graph reinforces the fact that the average price of gas nationwide is
between $2.60 and $2.90 per gallon. Moreover, gas prices are positively skewed since the
distribution runs off to the right; only two states (California and Washington) have gas
prices that are more than $3.00 per gallon.
Another useful visual representation of the data is an ogive, shown in Figure 2.B. The
ogive graphs the cumulative relative frequency distribution from Table 2.A. The ogive is
useful for approximating the “middle” price. If we draw a horizontal line on the ogive at
the 0.5 relative frequency mark, it intersects the plot at a point corresponding on the hori
zontal axis to a “middle price” of approximately $2.75. This indicates that gas stations in
approximately half of the states charged below this price and half charged above it.
Conceptual Review
LO 2.1 Summarize qualitative data by forming frequency distributions.
For qualitative data, a frequency distribution groups data into categories and re
cords the number of observations that fall into each category. A relative frequency
distribution shows the proportion (or the fraction) of observations in each category.
a. Construct a relative frequency distribution that a. Construct a relative frequency distribution that
summarizes the responses o f residents from Utah and summarizes the responses of the patrons. Briefly
Kentucky. Comment on the results. summarize your findings. What recommendations
b. Construct a bar chart tha t summarizes the results for would you make to the owner of the restaurant?
each state. b. Construct a pie chart and a bar chart for these data.
42. A survey conducted by CBS News asked parents about the Financial Resolution Frequency
professions they would want their children to pursue. The Saving more 328
results are summarized in the following table.
Paying down debt 257
Profession Parents' Preference Making more income 154
Doctor, banker, lawyer, or president 65% Spending less 133
internet mogul 13 Investing more 103
Humanitarian-aid worker 6 Saving for a large purchase 41
Athlete 9 Don't know 10
Movie star, rock star 2
a. Construct a relative frequency distribution for these data.
Other 5
What percent of the sample indicated that paying down
Source: Vanity Fair, D ecem ber 2009. debt was their top financial resolution?
a. Construct a bar chart and a pie chart for these data. b. Construct a bar chart.
b. How many parents wanted their children to become 46. A recent poll of 3,057 individuals asked: "What's the longest
athletes if the above results were based on 550 vacation you plan to take this summer?" The following
responses? relative frequency distribution summarizes the results.
43. The one-year return (in %) for 24 m utual funds is as follows: Response Relative Frequency
4.8 -16.8 9.0 6.5 8.2 5.3 A few long weekends 18%
1250 up to 1500 6 data set can be found on the text website and is labeled
Wealthiest Americans.
1500 up to 1750 10
1750 up to 2000 20 Name Age Net Worth ($ billions)
2000 up to 2250 15 William Gates III 53 50.0
2250 up to 2500 4 Warren Buffet 79 40.0
b. What is the most likely attendance range? How many a. What percent of the wealthiest people in America had
times was attendance less than 2000 people? net worth more than $20 billion?
c. What percent o f the tim e was attendance at least 1750 b. What percent o f the wealthiest people in America had
but less than 2000 people? What percent of the tim e was net w o rth between $10 billion and $20 billion?
attendance less than 1750 people? What percent o f the
c. Construct a stem-and-leaf diagram on age. Comment on
tim e was attendance 1750 or more?
the shape of the distribution and how it compares w ith
d. Construct a histogram. Comment on the shape of the the one for 2010 in Table 2.16.
distribution.
53. F IL E The price-to-earnings growth ratio, or PEG ratio, is
50. A researcher conducts a mileage economy test involving 80 the market's valuation o f a company relative to its earnings
cars. The frequency distribution describing average miles per prospects. A PEG ratio of 1 indicates tha t the stock's price
gallon (mpg) appears in the follow ing table. is in line w ith growth expectations. A PEG ratio less than
Average mpg Frequency 1 suggests that the stock o f the company is undervalued
(typical o f value stocks), whereas a PEG ratio greater than 1
15 up to 20 15
suggests the stock is overvalued (typical of growth stocks).
20 up to 25 30
The accompanying table shows a portion o f PEG ratios of
25 up to 30 15 companies listed on the Dow Jones Industrial Average; the
30 up to 35 10 complete data set can be found on the text website and is
35 up to 40 7 labeled DOW PEG Ratios.
40 up to 45 3
Company PEG Ratio
a. Construct the corresponding relative frequency, 3M (MMM) 1.4
cumulative frequency, and cumulative relative frequency Alcoa (AA) 0.9
distributions. ⋮ ⋮
b. How many o f the cars got less than 30 mpg? What Walt Disney (DIS) 1.2
percent of the cars got at least 20 but less than 25 mpg?
Source: w w w .fin a n c e .y a h o o , d a ta retrieved April 1 3 , 2011.
What percent o f the cars got less than 35 mpg? What
percent go t 35 mpg or more? Construct a stem-and-leaf diagram on the PEG ratio.
c. Construct a histogram. Comment on the shape of the Interpret your findings.
distribution. 54. The follow ing table lists the sale price and type o f 20 recently
51. The Wall Street Journal (August 2 8 , 2006) asked its readers: sold houses in New Jersey.
"Ideally, how many days a week, if any, would you work
Price Type Price Type
from home?" The follow ing relative frequency distribution
$305,000 Ranch $568,000 Colonial
summarizes the responses from 3,478 readers.
$450,000 Colonial $385,000 Other
Days Working from Home Relative Frequency
$389,000 Contemporary $310,000 Contemporary
0 0.12
$525,000 Other $450,000 Colonial
1 0.18
$300,000 Ranch $400,000 Other
2 0.30
$330,000 Contemporary $359,000 Ranch
3 0.15
$355,000 Contemporary $379,000 Ranch
4 0.07
$405,000 Colonial $509,000 Colonial
5 0.19
$365,000 Ranch $435,000 Colonial
Construct a pie chart and a bar chart to summarize the data. $415,000 Ranch $510,000 Other
a. Construct a pie chart and a bar chart on types of houses 56. The follow ing table lists the National Basketball Association's
sold in New Jersey. Interpret your findings. (NBA's) leading scorers, their average minutes per game
b. Construct a frequency distribution using seven classes, (MPG), and their average points per game (PPG) for 2008:
w ith the first class starting from $300,000. Player MPG PPG
c. Use a histogram and an ogive to summarize the data. D. Wade 38.6 30.2
55. A manager of a local retail store analyzes the relationship L. James 37.7 28.4
between advertising (in $100s) and sales (in $1,000s) by K. Bryant 36.1 26.8
reviewing the store's data for the previous six months. D. Nowitzki 37.3 25.9
Construct a scatterplot and com m ent on whether or not a
D. Granger 36.2 25.8
relationship exists.
K. Durant 39.0 25.3
Advertising (in $100s) Sales (in $1,000s)
C. Paul 38.5 22.8
20 15
C. Anthony 34.5 22.8
25 18
C. Bosh 38.0 22.7
30 20
B. Roy 37.2 22.6
22 16
Source: w w w .e s p n .c o m .
27 19
Construct and interpret a scatterplot o f PPG against MPG.
26 20
Does a relationship exist between the tw o variables?
CASE STUDIES
Data for Case Study 2.1 Net Revenue o f Nike in 2000 and 2009
Region 2000 2009
U.S. Region $4,732.1 $6,542.9
EMEA Regiona 2,350.9 5,512.2
Asia Pacific Region 955.1 3,322.0
Americas Region 550.2 1,284.7
Total = 8,588.3 Total = 16,661.8
Data for Case Study 2.2 Net Profit Margin for Clothing Retailers
Firm Net Profit Margin (in percent)
Abercrombie & Fitch 1.58
Aeropostale 10.64
Data for Case Study 2.3 Median Housing Prices by State, 2005
State Median
California $477,700
Hawaii 453,600
Mississippi 82,700
LEARNING OBJECTIVES
After reading
this chapter
you should
R
Measures
LO 3.1
LO 3.2
Calculate and interpret the arithmetic mean, the median, and the mode.
LO 3.4 Calculate and interpret the range, the mean absolute deviation, the
variance, the standard deviation, and the coefficient of variation.
LO 3.7 Calculate the mean and the variance for grouped data.
LO 3.8 Calculate and interpret the covariance and the correlation coefficient.
In Chapter 2 we learned how to summarize data by using tables and graphs so that we can extract
meaningful information. In this chapter w e focus on numerical descriptive measures. These mea
sures provide precise, objectively determined values tha t are easy to calculate, interpret, and co m
pare w ith one another. We first calculate several measures o f central location, which a tte m p t to find
a typical or central value for the data. In addition to analyzing the center, we need to know how the
data vary around the center. Measures o f spread or dispersion gauge the underlying variability o f the
data. We use measures o f central location and dispersion to introduce some popular applications, in
cluding the Sharpe ratio and the empirical rule. Finally, we discuss measures tha t examine the linear
relationship between tw o variables. These measures assess w hether tw o variables have a positive
linear relationship, a negative linear relationship, or no linear relationship.
Investment Decision
Rebecca J o h n s o n w o rk s as an in v e s tm e n t c o u n s e lo r a t a la rg e b a n k . R ecently, an in e x p e rie n c e d in v e s to r asked
J o h n s o n a b o u t c la rify in g s o m e d iffe re n c e s b e tw e e n t w o t o p - p e r fo r m in g m u tu a l fu n d s fr o m th e last d e c a d e :
V a n g u a rd 's P recious M e ta ls a n d M in in g fu n d (h e n c e fo rth , M etals) a n d F id e lity 's S tra te g ic In c o m e F und (h e n c e
fo r th , ln c o m e ).T h e in v e s to r s h o w s J o h n s o n th e re tu rn d a ta he has accessed o v e r th e In te rn e t, b u t th e in v e s to r has
tr o u b le in te r p r e tin g th e da ta . T able 3.1 s h o w s th e re tu rn d a ta fo r th e s e t w o m u tu a l fu n d s fo r th e years 2 0 0 0 -2 0 0 9 ;
th e d a ta , la b e le d F u n d R eturns, can also be fo u n d o n th e t e x t w e b s ite .
TABLE 3.1 Returns (in percent) for the Metals and the Income Funds, 2000-2009
1. D e te rm in e th e ty p ic a l re tu rn o f th e m u tu a l fu n d s .
2. E valuate th e in v e s tm e n t risk o f t h e m u tu a l fu n d s .
EXAMPLE 3.1
Let’s use the data in Table 3.1 in the introductory case to calculate and interpret the
mean return of the Metals fund and the mean return of the Income fund.
SOLUTION: Let’s start with the mean return for the Metals fund. We first add all the
returns and then divide by the number of returns as follows:
Similarly, we calculate the mean return for the Income fund as:
Thus, over the 10-year period 2000-2009, the mean return for the Metals fund was
greater than the average return for the Income fund, or equivalently, 24.65% >
8.51%. These arithmetic means represent typical annual returns resulting from
a one-year investment. Later we will discuss the geometric mean to describe the
annual return resulting from a multi-year investment.
All of us have calculated a mean before. What might be new for some of us is the notation
used to express the mean as a formula. For instance, when calculating the mean return for
the Metals fund, we let x 1 = – 7.34, x 2 = 18.33, and so on, and let n represent the number
of observations in the sample. So our calculation for the mean can be written as
The mean of the sample is referred to as (pronounced x-bar). Also, we can denote the
numerator of this formula using summation notation, which yields the following com
pact formula for the sample mean: We should also point out that if we had all
the return data for this mutual fund, instead of just the data for the past 10 years, then
we would have been able to calculate the population mean μ as where μ is
the Greek letter mu (pronounced as “mew”), and N is the number of observations in the
population.
THE A R I T H M E T IC ME A N
For sample values, x 1, x2, ., xn, the sample mean is computed as
For population values, x1, x2, . . . , xN, the population mean μ is computed as
The calculation method is identical for the sample mean and the population mean except
that the sample mean uses n observations and the population mean uses N observations,
where n < N. In later chapters we will refer to the population mean as a parameter and
the sample mean as a statistic. Since the population mean is generally unknown, we often
use the sample mean to estimate the population mean.
The arithmetic mean is used extensively in statistics. However, it can give a mislead
ing description of the center of the distribution in the presence of extremely small or large
values.
The arithmetic mean is the most commonly used measure of central location.
One weakness of this measure is that it is unduly influenced by outliers, that is,
extremely small or large values.
EXAMPLE 3.2
Seven people work at Acetech, a small technology firm in Seattle. Their salaries
over the past year are listed in Table 3.2. Compute the mean salary for this firm and
discuss whether it accurately indicates a typical value.
Title Salary
Administrative Assistant $40,000
Research Assistant 40,000
Computer Programmer 65,000
Senior Research Associate 90,000
Senior Sales Associate 145,000
Chief Financial Officer 150,000
President (and owner) 550,000
It is true that the mean salary for this firm is $154,286, but this value does not re
flect the typical salary at this firm. In fact, six of the seven employees earn less than
$154,286. This example highlights the main weakness of the mean, that is, it is very
sensitive to extreme observations (extremely large or extremely small values), or
outliers.
The M edian
Since the mean can be affected by outliers, we often also calculate the median as a
measure of central location. The median is the middle value of a data set. It divides the
data in half; an equal number of observations lie above and below the median. Many
government publications and other data sources publish both the mean and the median
in order to accurately portray a data set’s typical value. If the values of the mean and
the median differ significantly, then it is likely that the data set contains outliers. For
instance, in 2007 the United States Census Bureau determined that the median income
for American households was $46,326, whereas the mean income was $63,344. It is
well documented that a small number of households in the U.S. have income consider
ably higher than the typical American household income. As a result, these top-earning
households influence the mean by pushing its value significantly above the value of the
median.
THE M E D IA N
The median is the middle value of a data set. We arrange the data in ascending (or
descending) order and calculate the median as
• The middle value if the number of observations is odd, or
• The average of the two middle values if the number of observations is even.
The median is especially useful when outliers are present.
EXAMPLE 3.3
Use the data in Table 3.2 to calculate the median salary of employees at Acetech.
SOLUTION: In Table 3.2, the data are already arranged in ascending order. We
reproduce the salaries along with their relative positions.
Position: 1 2 3 4 5 6 7
Value: $40,000 40,000 65,000 90,000 145,000 150,000 550,000
Given seven salaries, the median occupies the 4th position. Thus, the median is
$90,000. Three salaries are less than $90,000 and three salaries are greater than
$90,000. As compared to the mean income of $154,286, the median in this case
better reflects the typical salary.
EXAMPLE 3.4
Use the data in Table 3.1 in the introductory case study to calculate and interpret the
median returns for the Metals and the Income funds.
SOLUTION: Let’s start with the median return for the Metals fund. We first arrange
the data in ascending order:
1
Position: 2 3 4 5 6 7 8 9 10
Value: -56.02 -7 .3 4 8.09 18.33 33.35 34.30 36.13 43.79 59.45 76.46
Given 10 observations, the median is the average of the values in the 5th and 6th oo-
sitions. These values are 33.35 and 34.30, so the median is
Over the period 2000-2009, the Metals fund had a median return of 33.83%, which
indicates that 5 years had returns less than 33.83% and 5 years had returns greater
than 33.83%. A comparison of the median return (33.83%) and the mean return
(24.65%) reveals a mean that is less than the median by almost 10 percentage
points, which means that the Metals data are affected by outliers. Thus, in order to
give a more transparent description of a data’s center, it is wise to report both the
mean and the median.
Similarly we can find the median for the Income fund as 7.34%. In this case, the
median return of 7.34% does not appear to deviate drastically from the mean return
of 8.51%. This is not surprising, since a casual inspection reveals that the relative
magnitude of outliers is weaker in the Income fund data.
Note that the mean and the median suggest that a typical annual return for the Metals
fund is much higher than the Income fund. Then why would anyone want to invest in the
Income fund? We will come back to this question later in this chapter, when we explore
the risk associated with these funds.
The M o de
The mode of a data set is the value that occurs most frequently. A data set can have
more than one mode, or even no mode. For instance, if we try to calculate the mode
return for either the Metals fund or the Income fund in Table 3.1, we see that no value
in either fund occurs more than once. Thus, there is no mode value for either fund. If
a data set has one mode, then we say it is unimodal. If two modes exist, then the data
set is bimodal; if three modes exist, then it is trimodal. Generally, the m ode’s value
as a measure of central location tends to diminish with data sets that have more than
three modes.
THE M O D E
The mode is the most frequently occurring value in a data set. A data set may have
no mode or more than one mode. The mode is the only meaningful measure of cen
tral location that can be used to summarize qualitative data.
EXAMPLE 3.5
Use the data in Table 3.2 to calculate the mode salary of employees at Acetech.
SOLUTION: The salary $40,000 is earned by two employees. Every other salary
occurs just once. So $40,000 is the mode salary. Just because a value occurs with the
most frequency does not guarantee that it best reflects the center of the data. It is true
that the mode salary at Acetech is $40,000, but most employees earn considerably
more than this amount.
s L L M S L M L L M
Comment on the data set and use the appropriate measure of central location that
best reflects the typical size of a sweatshirt.
SOLUTION: This data set is an example of ordinal data (qualitative data). Here, the
mode is the only relevant measure of central location. The mode size is L since it
appears 5 times as compared to S and M that appear 2 and 3 times, respectively. Often,
when examining issues relating to the demand for a product, such as replenishing
stock, the mode tends to be the most relevant measure of central location.
Metals Income
Mean 24.654 Mean 8.514
Standard Error 11.7414004 Standard Error 3.4997715
Median 33.825 Median 7.335
Mode #N/A Mode #N/A
Standard Deviation 37.1295681 Standard Deviation 11.067249
Sample Variance 1378.60483 Sample Variance 122.484
Kurtosis 1.668701 Kurtosis 2.3615757
Skewness – 1.0076169 Skewness 0.5602496
Range 132.48 Range 43.14
Minimum – 56.02 Minimum – 11.37
Maximum 76.46 Maximum 31.77
Sum 246.54 Sum 85.14
Count 10 Count 10
Note that there is no unique mode as no return in either fund occurs more than once;
Excel reports it as #NA. We would also like to comment on the numerical measures of
skewness and (excess) kurtosis that Excel reports, even though we have not discussed
their calculations. A skewness coefficient of zero indicates the data values are relatively
evenly distributed on both sides of the mean. A positive skewness coefficient implies that
extreme values are concentrated in the right tail of the distribution, pulling the mean up,
and the bulk of values lie to the left of the mean. Similarly, a negative skewness coeffi
cient implies that extreme values are concentrated in the left tail of the distribution, pull
ing the mean down, and the bulk of values lie to the right of the mean. We find that the
returns are negatively skewed (Skewness = – 1.0076) for the Metals fund and positively
skewed (Skewness = 0.5602) for the Income fund. A (excess) kurtosis coefficient mea
sures whether or not a distribution is more peaked with fatter tails (kurtosis > 0) or less
peaked with thinner tails (kurtosis < 0) than a normal distribution. As we discuss in later
chapters, the normal distribution, with the skewness and kurtosis coefficients of zero, is
a widely used distribution for statistical analysis. The kurtosis coefficient is positive for
the Metals as well as the Income returns, indicating that the return distributions are more
peaked and have fatter tails than the normal distribution. In other words, there is a higher
likelihood of outliers as compared to the normal distribution.
3. Given the follow ing observations from a population, corporation is its market capitalization (in billions o f dollars
calculate the mean, the median, and the mode. as of March 2 6 , 2010) and its total return to investors for the
year 2009. These data, labeled Largest Corporations, are
150 257 55 110 110 43 201 125 55 also available on the text website.
Mkt. Cap.
4. Given the follow ing observations from a population,
Company (in $ billions) Total Return
calculate the mean, the median, and the mode.
Walmart $209 – 2.7%
20 15 25 20 10 15 25 20 15 Exxon Mobil 314 – 12.6
6. The follow ing table shows the 10 highest-paid chief Cancer 36 Alzheimer's 5
executive officers o f the last decade. Heart Disease 35 Diabetes 3
Respiratory 14 HIV/AIDS 1
Compensation
Injury 8 Infant Death 1
Name Firm (in millions)
Stroke 7
Lawrence Ellison Oracle $1,835.7
Source: Massachusetts Department o f Health, Massachusetts Deaths 2007.
Barry Diller IAC, Expedia 1,142.9
Ray Irani Occidental Petroleum 857.1 Which measure o f central location is most useful in
making staffing decisions at Massachusetts' hospitals?
Steve Jobs Apple 748.8
Explain.
Richard Fairbank Capital One 568.5
9. One im portant statistic in baseball is a pitcher's earned
Angelo Mozilo Countrywide 528.6
run average, or ERA. This num ber represents the average
Eugene Isenberg Nabors Industries 518.0
num ber o f earned runs given up by the pitcher per nine
Terry Semel Yahoo 489.6
innings. The follow ing table lists a portion o f the ERAs for
Henry Silverman Cendant 481.2 pitchers playing for the New York Yankees and the Baltimore
William McGuire UnitedHealth Group 469.3 Orioles as o f July 2 2 , 2010; the complete data, labeled ERA,
PERCENTILES
In general, the p th percentile divides a data set into two parts:
• Approximately p percent of the observations have values less than the p th
percentile;
• Approximately (100 – p ) percent of the observations have values greater than
the p th percentile.
Calculating t h e p th percentile
A. First arrange the data in ascending order.
B. Locate the approximate position of the percentile by calculating Lp:
where Lp indicates the location of the desired percentile p and n is the sample size.
For the population percentile, replace n by N. We set p = 50 for the median as it is
the 50th percentile.
C. Once you find the value for Lp, observe whether or not Lp is an integer:
• If Lp is an integer, then Lp denotes the location of the pth percentile. For instance,
if L20 is equal to 2, then the 20th percentile is equal to the second observation in
the ordered data set.
• If Lp is not an integer, we need to interpolate between two observations to approxi
mate the desired percentile. So if L20 is equal to 2.25, then we need to interpolate
25% of the distance between the second and third observations in order to find the
20th percentile.
EXAMPLE 3.7
Consider the information presented in the introductory case of this chapter. Calcu
late and interpret the 25th and the 75th percentiles for the Metals fund.
Thus, 25% of the returns were less than 4.23% and 75% of the returns were greater
than 4.23%.
For the 75th percentile: So, the
75th percentile is located 25% of the distance between the eighth and ninth obser
vations; it is calculated as
Thus, 75% of the returns were less than 47.71% and 25% of the returns were greater
than 47.71%.
Earlier we calculated the median or the 50th percentile for the Metals fund and obtained
a value of 33.83%. When we calculate the 25th, the 50th, and the 75th percentiles for
a data set, we have effectively divided the data into four equal parts, or quarters. Thus,
the 25th percentile is also referred to as the first quartile (Q l), the 50th percentile is
referred to as the second quartile (Q2), and the 75th percentile is referred to as the third
quartile (Q3).
We can define other dividing lines that split the data into smaller parts:
• Quintiles divide the data set into fifths.
• Deciles divide the data set into tenths.
• Percentiles divide the data set into hundredths.
A box plot, also referred to as a box-and-whisker plot, is a convenient way to graphically
display the smallest value (S), the quartiles (Q l, Q2, and Q3), and the largest value (L)
of a data set. Box plots are particularly useful when comparing data sets; they are also an
effective tool for identifying outliers. Using our results from the Metals fund, Table 3.4
summarizes the five values that we will plot:
The values in Table 3.4 are often referred to as the five-number summary for the data set.
We follow these steps to construct a box plot and also to detect outliers.
A. Plot the five-number summary values in ascending order on the horizontal axis.
B. Draw a box encompassing the first and third quartiles.
C. Draw a dashed line in the box at the median.
D. To determine if a given observation is an outlier, first calculate the difference
between Q3 and Q1. This difference is called the interquartile range or IQR. The
IQR represents the middle half of the data. Draw a line (“whisker”) that extends
from Q 1 to the smallest data value that is not farther than 1.5 × IQR from Q 1.
Similarly, draw a line that extends from Q3 to the largest data value that is not
farther than 1.5 × IQR from Q3.
E. Use an asterisk to indicate points that are farther than 1.5 × IQR from the box.
These points are considered outliers.
EXAMPLE 3.8
Construct the box plot for the Metals fund.
SOLUTION: Based on the information in Table 3.4, we calculate the IQR as the
difference between Q3 and Q1 or 47.71% – 4.23% = 43.48%. We then calculate
1.5 × IQR or 1.5 × 43.48% = 65.22%. The distance between Q1 and the smallest
value, 4.23 – ( – 56.02%) = 60.25%, is within the limit of 65.22%; thus, the
line will extend to the smallest value of –56.02% on the left side of the box plot
(Figure 3.2). Similarly, the distance between the largest value and Q3, 76.46% –
47.71% = 28.75, is also well within the limit of 65.22%; here the line will extend to
the right up to the largest value of 76.46%. Given the criteria for constructing a box
plot, there are no outliers in this data set.
From this box plot we can quickly grasp several points concerning the distribu
tion of returns for the Metals fund. First, returns range from – 56.02% to 76.46%,
with about half being less than 33.83% and half being greater than 33.83%. We
make two further observations: (1) the median is off-center within the box, being
located to the right of center, and (2) the left whisker is longer than the right whis
ker. This indicates that the distribution is negatively skewed. If the median is left
of center and the right whisker is longer than the left whisker, then the distribu
tion is positively skewed. If the median is in the center of the box and the left and
right whiskers are equidistant from their respective quartiles, then the distribution is
symmetric.
EXERCISES 3.2
Concepts 16. Consider the return data (in percent) for the Income fund in
Table 3.1.
11. Calculate the 20th, 50th, and 80th percentiles for the
a. Calculate and interpret the 25th, 50th, and 75th
follow ing data set:
percentiles.
120 215 187 343 268 196 312
b. Construct a box plot. Are there any outliers?
12. Calculate the 20th, 40th, and 70th percentiles for the c. Is the distribution symmetric? If not, com m ent on its
follow ing data set: skewness.
– 300 – 257 – 325 – 234 – 297 – 362 – 255 17. A price-earnings ratio or P/E ratio is calculated as a
firm's share price compared to the income or pro fit earned
13. Consider the follow ing data set:
by the firm per share. Generally, a high P/E ratio suggests
12 9 27 15 58 35 21 32 22 tha t investors are expecting higher earnings grow th in the
a. Calculate and interpret the 25th, 50th, and 75th percentiles. future compared to companies w ith a lower P/E ratio. The
accompanying table shows a portion o f 26 companies that
b. Construct a box plot. Are there any outliers?
comprise the Dow Jones Industrial Average and their P/E
14. Consider the follow ing data set:
ratios as o f July 23, 2010 (at the tim e data were retrieved,
0.04 0.10 – 0.05 – 0.02 0.08 0.15 – 0.09 data on four firms were not available). The complete data,
labeled PE Ratio are available on the text website.
a. Calculate and interpret the 25th, 50th, and 75th percentiles.
b. Construct a box plot. Are there any outliers? Company P/E Ratio
3M (MMM) 17
Applications American Express (AXP) 22
15. Scores on the final in a statistics class are as follows. ⋮
75 25 75 62 80 85 80 99 90 60 Wal-Mart (WMT) 14
a. Calculate and interpret the 25th, 50th, and 75th percentiles. percentiles.
b. Construct a box plot. Are there any outliers? Is the b. Construct a box plot. Are there any outliers? Is the
distribution symmetric? If not, com m ent on its skewness. distribution symmetric? If not, com m ent on its skewness.
TABLE 3.5 End o f Year Holdings Given an Initial Investm ent o f $1,000
Year Return Value at the End of Year
2009 10 percent $1,000 + 1,000(0.10) = $1,100
2010 – 10 percent $1,100+ 1,100(– 0.10) = $990
F O R M U L A F O R THE G E O M E TR IC M E A N RETURN
For multiperiod returns R 1, R 2, ., Rn, the geometric mean return GR is computed
as
Let us revisit the above case where you invested $1,000 in a stock that had a 10% return
in 2009 and a – 10% return in 2010. The geometric mean is computed as
We interpret the geometric mean return as the annualized return, that you will earn from
a two-year investment period. Table 3.6 shows that with the computed annualized return
of – 0.5%, the end investment value is the same as shown in Table 3.5.
EXAMPLE 3.9
Use the data in Table 3.1 to calculate the geometric mean for the Metals and the
Income funds.
SO LU TIO N:
Metals Fund:
Income Fund:
Therefore, for the 10-year period, the annualized return for the Metals fund is
higher than that of the Income fund, 17.79% > 8.01%. However, the magnitude
of the difference is relatively smaller than that of the arithmetic means, which for
the Metals and Income funds are 24.65% and 8.51%, respectively. This shows that
the geometric mean is not as sensitive to extreme values as is the arithmetic mean.
The arithmetic mean for the Metals fund is unduly influenced by the extreme return
of 76.46% in 2009.
An issue that begs for explanation is the relevance of the arithmetic mean and the geomet
ric mean as summary measures for financial returns. Both means are relevant descriptive
measures for annual return; however, each has a different interpretation. The arithmetic
mean is appropriate for analyzing a one-year investment, whereas the geometric mean is
appropriate for analyzing a multi-year investment. In Example 3.9, the arithmetic mean
of 24.65% is the average annual return for summarizing returns with an investment ho
rizon of one year. The geometric mean of 17.79% is the average annual return when the
investment horizon is 10 years. For illustration, we can think of the arithmetic mean
as the relevant metric for an investor who is saving/investing to buy a house in about a
year’s time. The geometric mean is the relevant metric for an investor who is saving for
retirement.
F O R M U L A F O R T H E A V E R A G E G R O W T H RATE
For growth rates g 1, g2, ... , gn, the average growth rate Gg is computed as:
EXAMPLE 3.10
Table 3.7 shows sales for Adidas (in millions of €) for the years 2005 through 2009.
TABLE 3.7 Sales fo r Adidas (in m illions o f €), 2005–2009
Year 2005 2006 2007 2008 2009
Sales 6,636 10,084 10,299 10,799 10,381
Calculate the growth rates for 2005–2006, 2006–2007, 2007– 2008, and 2008– 2009
and use them to compute the average growth rate.
S O LU TIO N : The growth rates for Adidas for four years are computed as:
Therefore,
Sales for Adidas from 2005 to 2009 had an average growth rate of 11.84% per year.
There is a simpler way to compute the average growth rate when the underlying values of
the series are given. In the above example, it is cumbersome to first calculate the relevant
growth rates and then use them to compute the average growth rate.
A N A L T E R N A T I V E F O R M U L A F OR T H E A V E R A G E G R O W T H RATE
For observations x1, x2, ..,xn, the average growth rate Gg is computed as:
where n – 1 is the number of distinct growth rates. Note that only the first and last
observations are needed in the time series due to cancellations in the formula.
EXAMPLE 3.11
Calculate the average growth rate for Adidas directly from the sales data in
Table 3.7.
SO LU TIO N :Using the first and last observations from the time series consisting of
five observations, we calculate
EXERCISES 3.3
a. Calculate the growth rates for Year 1-Year 2, a. Calculate and interpret the arithm etic mean return.
Year 2-Year 3, and Year 3-Year 4.
b. Calculate and interpret the geometric mean return.
b. Calculate the average growth rate. c. How much money would you have accumulated by the
24. Consider the follow ing observations o f a tim e series: end o f 2009?
27. Suppose at the beginning o f 2005 you decide to invest a. Calculate the grow th rate for 2008–2009 and 2009–2010
$20,000 in Driehaus' Emerging Markets Growth mutual for each retailer.
fund. The follow ing table shows the returns for the years b. Calculate the average growth rate for each retailer.
2005-2009.
29. The follow ing table shows the total revenue (in billions of $)
Year Annual Return for Walmart Stores, Inc. and Target Corp. for the years
2005 0.2585 percent 2008-2010.
2006 0.2755 percent Year 2008 2009 2010
2007 0.2747 percent Walmart 379.8 404.3 408.2
2008 – 0.4702 percent Target 63.4 65.0 65.3
2009 0.7575 percent Source: Annual Reports o f Walmart Stores, Inc., and Target Corp.
Source: h t tp ://w w w .fin a n ce .ya h o o .c o m . a. Calculate the average growth rate for each firm.
a. Calculate and interpret the arithm etic mean return. b. Which firm had the higher grow th rate over the
b. Calculate and interpret the geometric mean return. 2008-2010 period?
c. How much money would you have accumulated by the 30. The follow ing table shows sales for Nike (in m illions o f $) for
end o f 2009? the years 2005 through 2009.
28. Home Depot and Lowe's are the tw o largest home Year 2005 2006 2007 2008 2009
improvem ent retailers in the U.S.The follow ing table shows Sales 13,740 14,955 16,326 18,627 19,176
the total revenue (in billions) for each retailer for the years
Source: Annual Reports o f Nike, Inc.
2008–2010.
a. Use the growth rates for 2005– 2006, 2006–2007,
Year Home Depot Lowe’s
2007–2008, and 2008-2009 to calculate the average
2008 $77.35 $48.28 growth rate.
2009 71.29 48.23 b. Calculate the average grow th rate directly from sales.
2010 66.18 47.22
Source: Annual Reports o f Home Depot, Inc., and Lowe's Companies Inc.
TABLE 3.8 Select Measures for the Metal and Income Funds, 2000-2009
Minimum Return Average Return Maximum Return
Metals fund -56.02% 24.65% 76.46%
Income fund -11.37% 8.51% 31.77%
We now discuss several measures of dispersion that gauge the variability of a data set.
Each measure is a numerical value that equals zero if all data values are identical, and
increases as data values become more diverse.
Range
The range is the simplest measure of dispersion; it is the difference between the maxi
mum and the minimum values in a data set.
EXAMPLE 3.12
Use the data in Table 3.8 to calculate the range for the Metals and the Income funds.
SOLUTION:
Metals fund: 76.46% – (– 56.02%) = 132.48%
Income fund: 31.77% – (– 11.37%) = 43.14%
The Metals fund has the higher value for the range, indicating that it has more
dispersion with respect to its minimum and maximum values.
The range is not considered a good measure of dispersion because it focuses solely on the
extreme values and ignores every other observation in the data set. While the interquartile
range, IQR = Q3 – Q1, discussed in Section 3.2, does not depend on the extreme values,
this measure still does not incorporate all the data.
THE M E A N ABSOLUTE D E V IA T IO N (M A D )
For sample values, x1, x 2, . . . , x n, the sample MAD is computed as
EXAMPLE 3.13
Use the data in Table 3.1 to calculate MAD for the Metals and the Income funds.
SOLUTION: We first compute the MAD for the Metals fund. The second column
in Table 3.9 shows differences from the sample mean, As mentioned
above, the sum of these differences equals zero (or a number very close to zero due
to rounding). The third column shows the absolute value of each deviation from the
mean. Summing these values yields the numerator for the MAD formula.
TABLE 3.9 MAD Calculations fo r the Metals Fund
THE VA R IA N C E A N D THE S TA N D A R D D E V IA T IO N
For sample values, x1, x2, . . . , xn, the sample variance s2 and the sample standard
deviation ,s are computed as
and
For population values, x1, x2, . . . , xN, the population variance σ 2 and the popula
tion standard deviation σ are computed as
and
Note: The sample variance uses n – 1 rather than n in the denominator; the reason
is discussed in Chapter 8.
EXAMPLE 3.14
Use the data in Table 3.1 to calculate the sample variance and the sample standard
deviation for the Metals and the Income funds. Express the answers in the correct
units of measurement.
SOLUTION: We will show the calculations for the Metals fund with the mean return
of 24.65 percent. The second column in Table 3.10 shows each return less the mean.
The third column shows the square of each deviation from the mean. Summing
these values yields the numerator for the sample variance formula.
TA B LE 3 .1 0 Sample Variance Calculation fo r the Metals Fund
Xi
Note that the units of measurement are squared. The sample standard deviation is
Based on all measures of dispersion discussed thus far, we can conclude that the
Income fund is less dispersed than the Metals fund. With financial data, standard
deviation tends to be the most common measure of risk. Therefore the investment
risk of the Income fund is lower than that of the Metals fund.
Some people prefer to use a shortcut formula for computing the variance.
S H O R T C U T F O R M U L A F OR T H E V A R I A N C E
We recommend that you use the shortcut formula to replicate the results of Example 3.14.
T H E C O E F F I C I E N T OF V A R I A T I O N ( C V )
Sample
Population
EXAMPLE 3.15
Calculate and interpret the coefficient of variation for the Metals and Income funds.
SOLUTION: We use the sample means and the standard deviations computed earlier.
Since 1.51 is greater than 1.30, we can conclude that the data for the Metals fund
has more relative dispersion than the Income fund.
Similarly, we can compute the population variance by using the VAR.P function in Excel.
Alternatively, we can compute the population variance by simply multiplying the sample
variance that Excel provides by
SYNOPSIS OF INTRODUCTORY CASE
EXERCISES 3.4
May 2010 26 81 a. Use Excel to calculate the mean and standard deviation
for m onthly rent.
June 2010 24 75
b. Use Excel to calculate the mean and standard deviation
Source: h ttp ://w w w .fin a n c e .y a h o o .c o m .
for square footage.
a. Calculate the sample variance and sample standard
c. Which variable has greater relative dispersion?
deviation for each firm's stock price.
38. FILE Go to the text website and access the data labeled
b. Which firm's stock price had greater variability as
Largest Corporations. It shows the Fortune 500 rankings
measured by standard deviation?
o f America's largest corporations for 2010. Next to each
c. Which firm's stock price had the greater relative
corporation are its market capitalization (in billions of
dispersion?
dollars as o f March 2 6 , 2010) and its total return to investors
37. FILE W hile the housing m arket is in recession and is not for the year 2009.
likely to emerge anytim e soon, real estate investm ent
a. Calculate the coefficient of variation for market
in college tow ns continues to promise good returns
capitalization.
(The Wall Street Journal, September 24,2010). Marcela
b. Calculate the coefficient of variation for total return.
Treisman works for an investm ent firm in M ichigan. Her
c. Which variable has greater relative dispersion?
assignment is to analyze the rental m arket in Ann Arbor,
Consider Table 3.11, which summarizes the mean and variance for the Metals and Income
funds.
TABLE 3 .1 1 Mean-Variance Analysis o f Two Mutual Funds, 2000–2009
It is true that the Metals fund provided an investor with a higher reward over the 10-
year period, but this same investor encountered considerable risk compared to an in
vestor who invested in the Income fund. Table 3.11 shows that the variance of the
Metals (1,378.61(%)2) fund is significantly greater than the variance of the Income fund
(122.48(%)2). If we look back at Table 3.1 and focus on the Metals fund, we see returns
far above the average return of 24.65% (for example, 59.45% and 76.46%), but also
returns far below the average return of 24.65% (for example, – 7.34% and – 56.02%).
Repeating this same analysis for the Income fund, the returns are far closer to the average
return of 8.51%; thus, the Income fund provided a lower return, but also far less risk.
A discussion of mean-variance analysis seems almost incomplete without mention
of the Sharpe ratio. Nobel Laureate William Sharpe developed what he originally re
ferred to as the “reward-to-variability” ratio. However, academics and finance profession
als prefer to call it the “Sharpe ratio.” The Sharpe ratio is used to characterize how well
the return of an asset compensates for the risk that the investor takes. Investors are often
advised to pick investments that have high Sharpe ratios.
The Sharpe ratio is defined with the reward specified in terms of the population mean
and the variability specified in terms of the population variance. However, we often com
pute the Sharpe ratio in terms of the sample mean and sample variance, where the return
is usually expressed as a percent and not a decimal.
where is the mean return for the investment, is the mean return for a risk-free
asset such as a Treasury bill (T-bill), and s, is the standard deviation for the investment.
The numerator of the Sharpe ratio measures the extra reward that investors receive for
the added risk taken— this difference is often called excess return. The higher the Sharpe
ratio, the better the investment compensates its investors for risk.
EXAMPLE 3.16
Calculate and interpret the Sharpe ratios for the Metals and Income funds given that
the return on a 1-year T-bill is 4%.
SOLUTION: Since the return on a 1-year T-bill is 4%, Plugging in the values
of the relevant means and standard deviations into the Sharpe ratio yields:
39. Consider the follow ing data for tw o investments, A and B: 2005 44.60 15.41
40. Consider the follow ing data for tw o investments, A and B: a. Which fund had the higher average return?
b. Which fund was riskier over this tim e period?
Investment A:
c. Given a risk-free rate o f 3%, which fund has the
Investment B:
higher Sharpe ratio? W hat does this ratio imply?
a. Which investment provides the higher return? Which
43 The fo llo w in g table shows th e annual returns (in
investment provides the least risk? Explain.
percent) and sum m ary measures fo r the Fidelity Latin
b. Given a risk-free rate o f 1.4%, calculate the Sharpe ratio America Fund and the Fidelity Canada Fund from 2005
for each investment. Which investment provides the thro u g h 2009.
higher reward per unit o f risk? Explain.
Year Latin America Canada
41. Consider the follow ing returns for tw o investments, A and B:
2005 55.17 27.89
Investment 1: 2% 8% – 4% 6%
2006 44.33 15.04
Investment 2: 6% 12% – 8% 10% 2007 43.71 35.02
a. Which investment provides the higher return? 2008 – 54.64 – 42.64
b. Which investment provides the least risk? 2009 91.60 39.63
c. Given a risk-free rate o f 1.2%, calculate the Sharpe ratio Source: h ttp ://w w w .fin a n ce .ya h o o .c o m .
for each investment. Which investm ent has performed a. Which fund had the higher average return?
better? Explain.
b. Which fund was riskier over this tim e period?
c. Given a risk-free rate o f 3%, which fund has the higher
Applications
Sharpe ratio? What does this ratio imply?
42. The follow ing table shows the annual returns (in percent)
and summary measures for the Vanguard Energy Fund and
the Vanguard Health Care Fund from 2005 through 2009.
Chebyshev's Theorem
As we will see in more detail in later chapters, it is important to be able to use the standard
deviation to make statements about the proportion of observations that fall within certain
intervals. Fortunately, a Russian mathematician Pavroty Chebyshev (1821– 1894) found
bounds for the proportion of the data that lie within a specified number of standard devia
tions from the mean.
CHEBYSHEV'S THEOREM
For any data set, the proportion of observations that lie within k standard deviations
from the mean is at least 1 – l / k2, where k is any number greater than 1.
EXAMPLE 3.17
A large lecture class has 280 students. The professor has announced that the mean
score on an exam is 74 with a standard deviation of 8. At least how many students
scored within 58 and 90?
The application of Chebyshev’s Theorem results in conservative bounds for the percent
age of observations falling in a particular interval. The actual percentage of observations
lying in the interval may in fact be much larger.
T H E E M P I R I C A L RULE
Given a sample mean a sample standard deviation s, and a relatively symmetric
and bell-shaped distribution:
• Approximately 68% of all observations fall in the interval
• Approximately 95% of all observations fall in the interval and
• Almost all observations fall in the interval
FIGURE 3.4 Graphical description o f the empirical rule
EXAMPLE 3.18
Let’s revisit Example 3.17 regarding a large lecture class with 280 students with a
mean score of 74 and a standard deviation of 8. Assume that the distribution is sym
metric and bell-shaped.
a. Approximately how many students scored within 58 and 90?
b. Approximately how many students scored more than 90?
SOLUTION:
a. As shown in Example 3.17, the score 58 is two standard deviations below
the mean while the score 90 is two standard deviations above the mean. The
empirical rule states that approximately 95% of the observations fall within
two standard deviations of the mean. Therefore about 95% of 280 students, or
0.95(280) = 266 students, scored within 58 and 90.
b. We know that the score 90 is two standard deviations above the mean. Since
approximately 95% of the observations fall within two standard deviations of
the mean, we can infer that 5% of the observations fall outside the interval.
Therefore, about half of 5%, or 2.5%, of 280 students scored above 90. Equiv
alently, about 7 students (0.025 × 280) scored above 90 on the exam. If the
professor uses a cutoff score above 90 for an A, then only seven students in the
class are expected to get an A.
The main difference between Chebyshev’s Theorem and the empirical rule is that
Chebyshev’s Theorem applies to all data sets whereas the empirical rule is appropriate
when the distribution is symmetric and bell-shaped.
The empirical rule also provides us with a rough approximation for the standard
deviation of the data if we only have information on the range of the data. Since the
empirical rule tells us that 95% of the observations fall within two standard deviations of
the mean, the resulting interval encompasses approximately 4 × 5 of the observations.
Thus, the range = 4 × s, or analogously, Keep in mind that this formula pro
vides only a rough estimate for the sample standard deviation.
EXERCISES 3.6
Concepts 51. Data w ith 250 observations are drawn from a relatively
symmetric and bell-shaped distribution w ith a mean o f 50
44. A data set has a mean o f 80 and a standard deviation o f 5.
and a range of 48.
a. Using Chebyshev's Theorem, w hat percentage o f the
a. What is the rough estimate of the standard deviation?
observations fall between 70 and 90?
b. Approximately how many observations are more than 74?
b. Using Chebyshev's Theorem, w hat percentage o f the
observations fall between 65 and 95?
Applications
45. A data set has a mean of 1500 and a standard deviation of
52. A sample of the salaries of assistant professors on the
100 .
business faculty at a local university revealed a mean income
a. Using Chebyshev's Theorem, w hat percentage of the
of $72,000 w ith a standard deviation o f $3,000.
observations fall between 1300 and 1700?
a. Using Chebyshev's Theorem, w hat percentage of the
b. Using Chebyshev's Theorem, w hat percentage o f the
faculty earns at least $66,000 but no more than $78,000?
observations fall between 1100 and 1900?
b. Using Chebyshev's Theorem, what percentage of the
46. A data set has a mean of 500 and a standard deviation faculty earns at least $63,000 but no more than $81,000?
of 25.
53. The historical returns on a portfolio had an average return of
a. Using Chebyshev's Theorem, find the interval that
8 percent and a standard deviation of 12 percent. Assume that
encompasses at least 75% of the data.
returns on this portfolio follow a bell-shaped distribution.
b. Using Chebyshev's Theorem, find the interval that
a. What percentage o f returns were greater than 20 percent?
encompasses at least 89% of the data.
b. What percentage o f returns were below -1 6 percent?
47. Data are drawn from a relatively symmetric and bell-shaped
54. It is often assumed that IQ scores follow a bell-shaped
distribution w ith a mean o f 20 and a standard deviation of 2.
distribution w ith a mean of 100 and a standard deviation of 16.
a. What percentage of the observations fall between 18
a. What percentage of scores are between 84 and 116?
and 22?
b. What percentage of scores are less than 68?
b. What percentage of the observations fall between 16
c. What percentage of scores are more than 116?
and 24?
c. What percentage o f the observations are less than 16? 55. An investment strategy has an expected return of 8 percent
and a standard deviation o f 6 percent. Assume investment
48. Consider a symmetric and bell-shaped distribution w ith a
returns are bell shaped.
mean of 750 and a standard deviation of 50. There are 500
a. How likely is it to earn a return between 2 percent and
observations in the data set.
14 percent?
a. What percentage o f the observations are less than 700?
b. How likely is it to earn a return greater than 14 percent?
b. Approximately how many observations are less than 700?
c. How likely is it to earn a return below –4 percent?
49. Data are drawn from a symmetric and bell-shaped
56. Average talk tim e between charges of a given cell phone
distribution w ith a mean o f 25 and a standard deviation of 4.
is advertised as 4 hours. Let the standard deviation be
There are 1000 observations in the data set.
0.8 hours.
a. What percentage of the observations are less than 33?
a. Use Chebyshev's Theorem to approximate the
b. Approximately how many observations are less than 33?
proportion of cell phones tha t w ill have talk tim e
50. Data are drawn from a relatively symmetric and bell-shaped between 2.4 hours and 5.6 hours.
distribution w ith a mean o f 5 and a range o f 10.
b. Assume a bell-shaped distribution to approximate
a. What is the rough estimate of the standard deviation? the proportion o f cell phones that will have talk tim e
b. W hat percentage of the observations are positive? between 2.4 hours and 5.6 hours.
c. W hat percentage of the observations are not positive?
C A L C U L A T IN G THE M E A N A N D THE V A R IA NC E
F OR A F R E Q U E N C Y D I S T R I B U T I O N
Sample: Population:
Mean: Mean:
Variance: Variance:
where m, and f are the midpoint and the frequency of the ith class, respectively. The
standard deviation is the positive square root of the variance.
Note that by aggregating, some of the data information is lost. Therefore, unlike in the
case of raw data, we can only compute approximate values of the summary measures with
grouped data.
EXAMPLE 3.19
Recall the frequency distribution of house prices we constructed in Chapter 2.
500 up to 600 14
600 up to 700 5
700 up to 800 2
SOLUTION: Table 3.12 shows the frequencies fi and the midpoint of each class mi in
the second and third columns, respectively.
TABLE 3.12 The Sample Mean and the Sample Variance Calculation fo r Grouped Data
Class (in $1,000s) fi mi mifi
300 up to 400 4 350 1,400 (350 - 522)2 × 4 = 118,336
400 up to 500 11 450 4,950 (450 - 522)2 × 11 = 57,024
500 up to 600 14 550 7,700 (550 – 522)2 × 14 = 10,976
600 up to 700 5 650 3,250 (650 – 522)2 × 5 = 81,920
700 up to 800 2 750 1,500 (750 – 522)2 × 2 = 103,968
Total 36 18,800 372,224
a. For the mean, we multiply each class’s midpoint by its respective frequency, as
shown in the fourth column of Table 3.12. Finally, we sum the fourth column
and divide by the sample size. Or,
The standard deviation is simply the positive square root of the sample variance, or,
The standard deviation is $103.13.
As in the case of raw data, some people prefer to use a shortcut formula for computing
the variance.
S H O R T C U T F O R M U L A F OR T H E V A R I A N C E
We recommend that you use the shortcut formula to replicate the results of Example 3.19.
Many times the data from secondary sources are distributed in the form o f a
relative frequency distribution rather than a frequency distribution. In order to use
the formulas for the mean and variance for grouped data, first convert the relative
frequency distribution into a frequency distribution, as discussed in Section 2.2 of
Chapter 2.
The sample mean calculated with grouped data can be thought of as a weighted mean
where the relative frequency f /i n is treated as a weight for the midpoint. The more general
formula for the weighted mean is given below.
THE W E IG H T E D ME AN
Let w1, w2, . . . , wn, denote the weights of the sample observations x h x2, . . . , x n such
that w 1 + w2 + ⋯ + wn = 1. The weighted mean for the sample is computed as
EXAMPLE 3.20
A student scores 60 on Exam 1, 70 on Exam 2, and 80 on Exam 3. What is the stu
dent’s average score for the course if Exams 1, 2, and 3 are worth 25%, 25%, and
50% of the grade, respectively?
a. Calculate the sample mean. Date Price Per Share Number of Shares
b. Calculate the sample variance and the sample standard January 2009 $19.58 70
deviation. July 2009 $24.06 80
59. The follow ing relative frequency distribution was December 2009 $29.54 50
constructed from a population of 200. Calculate the
Calculate the average price per share at which the investor
population mean, the population variance, and the
bought these shares.
population standard deviation.
64. A researcher conducts a mileage economy test involving
Class Relative Frequency
80 cars. The frequency distribution describing average
– 20 up to – 10 0.35
miles per gallon (mpg) appears below.
– 10 up to 0 0.25
Average MPG Frequency
0 up to 10 0.40
15 up to 20 15
10 up to 20 0.05
20 up to 25 30
60. The follow ing relative frequency distribution was 25 up to 30 15
constructed from a sample o f 50. Calculate the sample mean, 30 up to 35 10
the sample variance, and the sample standard deviation. 35 up to 40 7
Class Relative Frequency 40 up to 45 3
0 up to 2 0.34
a. Calculate the mean mpg.
2 up to 4 0.20
b. Calculate the variance and the standard deviation.
4 up to 6 0.40
65. The Boston Security Analysts Society, Inc. (BSAS) is a nonprofit
6 up to 8 0.06
association that serves as a forum for the exchange o f ideas
for the investment community. Suppose the ages o f its
Applications members are based on the follow ing frequency distribution.
61. Fifty cities provided inform ation on vacancy rates (in percent)
Age Frequency
in local apartments in the follow ing frequency distribution.
21-31 11
Vacancy Rate (in percent) Frequency
32-42 44
0 up to 3 5
43-53 26
3 up to 6 5
54-64 7
6 up to 9 10
a. Calculate the mean age.
9 up to 12 20
b. Calculate the sample variance and the sample standard
12 up to 15 10
deviation.
66. The National Sporting Goods Association (NSGA) conducted 67. You score 90 on the midterm, 60 on the final, and 80 on the
a survey of the ages o f people that purchased athletic class project. What is your average score if the m idterm is
footwear in 2009. The ages are summarized in the following w orth 30%, the final is w orth 50% and the class project is
relative frequency distribution. w orth 20%?
Age of Purchaser Percent 68. An investor bought common stock o f Dell Inc. Corporation
Under 14 years old 19 on three occasions at the follow ing prices.
THE C OV A R IA N C E
For values (x 1, y 1), (x 2, y2), . . . , (x n, yn), the sample covariance Sxv is computed as
For values (x1, y1), (x2, y2), ..., (xN, yN), the population covariance σ xy is computed
as
Note: As in the case of the sample variance, the sample covariance uses n — 1 rather
than n in the denominator.
The correlation coefficient is unit free since the units in the numerator cancel with those
in the denominator. The value of the correlation coefficient falls between – 1 and 1. A
perfect positive relationship exists if it equals 1, and a perfect negative relationship ex
ists if it equals – 1. Other values for the correlation coefficient must be interpreted with
reference to – 1, 0, or 1. For instance, a correlation coefficient equal to – 0.80 indicates
a strong negative relationship, whereas a correlation coefficient equal to 0.12 indicates a
weak positive relationship.
EXAMPLE 3.21
Calculate the covariance and the correlation coefficient for the Metals (x ) and
Income (y) funds. Interpret these values. Recall that sx = 37.13,
and.sv = 11.07.
SOLUTION: As a first step, Figure 3.5 shows a scatterplot of the return data for
the Metals and Income funds. It appears that there is a positive linear relationship
between the two fund returns.
FIGURE 3.5 Scatterplot o f return data for the Metals and Income funds
Table 3.13 shows the return data for each fund in the first two columns. The third
column shows the product of differences from the mean.
Summing the values in the third column yields the numerator for the covariance
formula. Thus, we calculate the covariance as:
T A B L E 3 .1 3 Covariance Calculation for the Metals and Income Funds
Xi yi
– 7.34 4.07 (– 7.34 – 24.65K4.07 – 8.51) = 142.04
18.33 6.52 (18.33 – 24.65)(6.52 – 8.51) = 12.58
⋮ ⋮ ⋮
76.46 31.77 (76.46 – 24.651(31.77 – 8.51) = 1,205.10
Total = 3,165.55
The covariance of 351.73 indicates that the variables have a positive linear relation
ship. In other words, on average, when one fund’s return is above its mean, the other
fund’s return is above its mean, and vice versa. The covariance is used to compute
the correlation coefficient as:
D. In order to calculate the sample correlation coefficient, choose Formulas > Insert
Function > CORREL. This command is appropriate for both sample and popula
tion data. Select the data as you did in part C. Excel returns the value 0.86, again
matching the value that we calculated manually.
EXERCISES 3.8
Concepts a. Calculate the covariance. W hat kind of linear relationship
exists?
69. Consider the follow ing sample data:
b. Calculate the coefficient of correlation. Comment on the
12 18 20 22 25
strength of the linear relationship.
15 20 25 22 27
73. The director o f graduate admissions at a local university is
a. Calculate the covariance between the variables. analyzing the relationship between scores on the Graduate
b. Calculate and interpret the correlation coefficient. Record Examination (GRE) and subsequent performance in
graduate school, as measured by a student's grade point
70. Consider the follow ing sample data:
average (GPA). She uses a sample of 10 students who
–2 0 3 4 7
graduated w ithin the past five years.
–2 –3 –8 –9 – 10
GRE GPA
a. Calculate the covariance between the variables. 1500 3.4
b. Calculate and interpret the correlation coefficient. 1400 3.5
1000 3.0
Applications
1050 2.9
71. The follow ing table shows the annual returns (in percent) for
1100 3.0
th e T-Rowe Price's Value and International Stock funds for the
1250 3.3
tim e period 2005-2009.
800 2.7
Year Value Fund International Fund
850 2.8
2005 6.30 16.27
950 3.2
2006 19.75 19.26
1350 3.3
2007 0.75 13.43
2008 – 39.76 – 48.02 a. Calculate and interpret the covariance.
2009 37.15 52.20 b. Calculate and interpret the coefficient of correlation. Does
an applicant's GRE score seem to be a good indicator of
a. Calculate and interpret the covariance between returns.
subsequent performance in graduate school?
b. Calculate and interpret the coefficient of correlation.
74. A social scientist wants to analyze the relationship between
72. In an attem pt to determine w hether a linear relationship educational attainm ent and salary. He collects the following
exists between the price o f a home and the number o f days sample o f data where "Education" refers to years o f higher
it takes to sell the home, a real estate agent collected the education and "Salary" is the person's annual salary in
follow ing data from recent sales in his city. thousands of dollars.
Price Days to Sell Price Days to Sell Education 3 4 6 2 5 4 8 0
(in $1,000s) Home (in $1,000s) Home
Salary $40 53 60 35 55 50 80 35
265 136 430 145
a. Calculate the covariance. W hat kind o f linear relationship
225 125 515 121
exists?
160 120 180 122
b. Calculate the coefficient o f correlation. Comment on the
325 140 423 145
strength of the linear relationship.
Many environmental groups and politicians are suggesting a return to the federal 55-mile-
per-hour (mph) speed limit on America’s highways. They argue that not only will a lower
national speed limit reduce greenhouse emissions, it will also increase traffic safety.
Cameron Grinnell believes that a lower speed limit will not increase traffic safety. He
believes that traffic safety is based on the variability of the speeds with which people are
driving, rather than the average speed. The person who drives 20 mph below the pace of
traffic is often as much a safety menace as the speeder. Cameron gathers the speeds of 40
cars from a highway with a speed limit of 55 mph (Highway 1) and the speeds of 40 cars
from a highway with a speed limit of 65 mph (Highway 2).
A portion of the data is shown in Table 3.14; the complete
data, labeled Highway Speeds, are available on the text
website.
Recently, many concerned citizens have lobbied for a return to the federal 55-mile-per- Sample
hour (mph) speed limit on America’s highways. The reduction may lower gas emissions
and save consumers on gasoline costs, but whether it will increase traffic safety is not Report—
clear. Many researchers believe that traffic safety is based on the variability of the speed
rather than the average speed with which people are driving— the more variability in Analyzing
speed, the more dangerous the roads. Is there less variability in speed on a highway with
a 55-mph speed limit as opposed to a 65-mph speed limit? Speed Limits
To compare average speeds, as well as the variability of speeds on highways, the
speeds of 40 cars were recorded on a highway with a 55-mph speed limit (Highway 1)
and a highway with a 65-mph speed limit (Highway 2). Table 3.A shows the most relevant
descriptive measures for the analysis.
Highway 1 Highway 2
(55-mph speed limit) (65-mph speed limit)
Mean 57 66
Median 56 66
Mode 50 70
Minimum 45 60
Maximum 74 70
Standard deviation 7.0 3.0
Coefficient of variation 0.12 0.05
Number of Cars 40 40
The average speed of a car on Highway 1 was 57 mph, as opposed to 66 mph on Highway
2. On Highway 1, half of the 40 cars drove faster than 56 mph and half drove slower than
56 mph, as measured by the median; the median for Highway 2 was 66 mph. The mode
shows that the most common speeds on Highway 1 and Highway 2 were 50 mph and 70
mph, respectively. Based on each measure of central location, Highway 2 experiences
higher speeds as compared to Highway 1.
While measures of central location typically represent where the data clusters, these
measures do not relay information about the variability in the data. Given the minimum
and maximum speeds of 45 mph and 74 mph, respectively, the range of speeds is
29 mph for Highway 1 as compared to a range of just 10 mph for Highway 2. Generally,
standard deviation is a more credible measure of dispersion, since range is based entirely
on the minimum and the maximum values. The standard deviation for Highway 1 is
substantially greater than the standard deviation for Highway 2 (7.0 mph > 3.0 mph).
Therefore, the speeds on Highway 1 are more variable than the speeds on Highway 2.
Even adjusting for differences in the magnitudes of the means by calculating the coef
ficient of variation, the speeds on Highway 1 are still more dispersed than Highway 2
(0.12 > 0.05).
On average, it is true that the speeds on Highway 2 are higher than the speeds on
Highway 1; however, the variability of speeds is greater on Highway 1. If traffic safety
improves when the variability of speeds declines, then the data suggest that a return to a
federal 55-mph speed limit may not enhance the well-being of highway travelers.
Conceptual Review
LO 3.1 Calculate and interpret the arithmetic mean, the median, and the mode.
The sample mean. and the population mean. are the arithmetic
averages of the data set. The mean is the most widely used measure of central location.
One weakness of the mean is that it is unduly influenced by outliers— extremely small
or large values.
The median is the middle value of a data set and is especially useful when outliers
are present or when the distribution is skewed. We arrange the data in ascending (or
descending) order and find the median as
• The middle value if the number of observations is odd, or
• The average of the two middle values if the number of observations is even.
The mode is the value in the data set that occurs with the most frequency. A data set
may have no mode or more than one mode. If the data are qualitative, then the mode is
the only meaningful measure of central location.
LO 3.3 Calculate and interpret a geometric mean return and an average growth rate.
The geometric mean is the multiplicative average of a data set. In general, the geo
metric mean is smaller than the arithmetic mean and is less sensitive to outliers. The
geometric mean is relevant when summarizing financial returns over several years.
For multiperiod returns R 1, R2, . . . , R n , the geometric mean return is computed as
where n is the number of multiperiod
returns.
The geometric mean is also used when summarizing average growth rates.
For growth rates g 1 g2, . . . , gn, the average growth rate is computed as
where n is the number of multiperiod
growth rates. When the underlying values of the series are given, there is a simpler way
to compute the average growth rate. For observations x 1, x2, . . . , xn, the average growth
rate is computed as
LO 3.4 Calculate and interpret the range, the mean absolute deviation, the variance, the
standard deviation, and the coefficient of variation.
The range is the difference between the maximum and the minimum values in a data set.
The mean absolute deviation (MAD) is an average of the absolute differences be
tween the observations and the mean of a data set. The sample MAD and the popula
tion MAD are computed as respectively.
The variance and standard deviation, which are based on squared differences from
the mean, are the two most widely used measures of dispersion. The sample variance s2
and the sample standard deviation s are computed as and
respectively. The population variance σ 2 and the population standard devia
tion σ are computed as respectively. Variance
squares the original units of measurement; by calculating the standard deviation, we
return to the original units of measurement.
The coefficient of variation C V is a relative measure of dispersion. The CV allows
comparisons of variability between data sets with different means or different units of
measurement. The sample CV and the population CV are computed as and
respectively.
of return and evaluate this rate of return in terms of its reward (mean) and risk (variance).
In general, investments with higher average returns are also associated with higher risk.
The Sharpe ratio measures extra reward per unit of risk. The Sharpe ratio for an
investment, I , is com puted as denotes the return on a risk-free asset.
The higher the Sharpe ratio, the better the investment compensates its investors for
risk.
LO 3.8 Calculate and interpret the covariance and the correlation coefficient.
The covariance and the correlation coefficient are measures that assess the existence
and strength of a linear relationship between two variables, x and y.
The sample covariance sxy and the population covariance σxy are computed as sxy =
respectively.
The sample correlation coefficient rxy and the population correlation coefficient ρxy are
computed as respectively.
S o u rce : The Boston Globe, July 3 1 , 2 0 1 0 . Source: A nnual Reports fo r Gap, In c, and A m erican Eagle O utfitters, Inc.
Calculate th e m ean, th e m edian, and th e m ode. W hich a. Calculate th e average g ro w th rate fo r each firm .
m easure o f central te n d e n c y best reflects typ ica l sales? b. W hich firm had th e h ig h e r g ro w th rate over th e
Explain. 2 0 0 8 -2 0 1 0 period?
79. M o n th ly stock prices fo r tw o c o m p e tin g firm s are Year G rowth Fund Fund o f Boston
as follow s.
2002 − 26.43 − 8.42
M onth Firm A Firm B 2003 24.71 17.40
January $28 $21 2004 4.80 8.32
February 31 24 2005 5.50 4.04
March 32 24 2006 5.23 12.25
April 35 27 Source: http://www.finance.yahoo.com.
May 34 25 a. Calculate and in te rp re t th e covariance.
June 28 20 b. Calculate th e co rrelation co efficient. C o m m e n t on th e
s tre n g th o f th e linear relationship.
a. Calculate th e sam ple m ean, th e sam ple variance,
83. A m anager o f a local retail store analyzes th e relatio n sh ip
and th e sam ple standard d e v ia tio n fo r each firm's
stock price. be tw e en a d ve rtisin g and sales by review ing th e store's data
fo r th e previous six m onths.
b. W hich firm had th e h ig h e r stock price over th e tim e
period? Advertising (in $100s) Sales (in $1,0 0 0 s)
c. W hich firm 's stock price had greater va ria b ility as 20 15
m easured by standard deviation? W hich firm 's stock 25 18
price had th e gre a te r relative dispersion? 30 20
80. The m anager a t a w a te r park constru cte d th e fo llo w in g 22 16
freq u e n cy d is trib u tio n to sum m arize atte nd a n ce fo r 60 days
27 19
in July and August.
26 20
Attendance Frequency
a. Calculate th e mean o f ad ve rtisin g and th e m ean o f sales.
1,000 up to 1,250 5
b. Calculate th e standard d e via tion o f ad ve rtisin g and th e
1,250 up to 1,500 6
standard d e via tio n o f sales.
1,500 up to 1,750 10
c. Calculate and in te rp re t th e covariance betw een
1,750 up to 2,000 20 a d ve rtisin g and sales.
2,000 up to 2,250 15 d. Calculate and in te rp re t th e co rrelation coefficient.
2,250 up to 2,500 4
84. F IL E A n eco n om ist wishes to sum m arize sam ple data
fro m 26 m e tro p o lita n areas in th e U.S. The fo llo w in g ta b le
a. Calculate th e mean attendance.
lists each area's 2010– 2011 m edian incom e as w e ll as th e
b. Calculate th e variance and th e standard d e via tion .
m o n th ly u n e m p lo y m e n t rate and average co nsum er d e b t fo r
81. The N ational S p o rtin g Goods Association (NSGA) August 2010; th e co m p le te data set can be fo u n d on th e te x t
co n d ucte d a survey o f th e ages o f in d ivid u als th a t w e b site and is labeled D e b t P a y m e n t s .
purchased ska teboarding fo otw ear. The ages o f this
M etropolitan Income
survey are sum m arized in th e fo llo w in g relative freq u e n cy
Area (in $1,000s) Unem ploym ent Debt
d is trib u tio n .
Washington, D.C. $103.50 6.3% $1,285
Age of User Percent Seattle 81.70 8.5 1,135
Under 14 years old 35 ⋮ ⋮ ⋮ ⋮
CASE STUDIES
Adidas’ Online Annual Report provides total revenues (in millions of €) for the Asian and
Latin American regions for the years 2005 through 2009 as follows:
Adidas Revenues in Asia and Latin America (in millions of €)
2005 2006 2007 2008 2009
Asia 1,523 2,020 2,254 2,662 2,614
Latin America 319 499 657 893 1,006
25 200 180
Source: w w w .zillo w .co m .
E v e r y d a y w e m a k e c h o ic e s a b o u t issues c o n c e r n i n g u n c e r t a i n e v e n t s . U s u a lly , w e e i t h e r i m p l i c i t l y
t h e p a p e r , w a t c h t h e n e w s , o r c h e c k t h e I n t e r n e t t o d e t e r m i n e t h e lik e li h o o d o f rain a n d w h e t h e r
w e s h o u l d c a r r y a n u m b r e l l a . R e ta ile r s s t r e n g t h e n t h e i r sales f o r c e b e f o r e t h e e n d - o f - y e a r h o l i d a y
s e a s o n in a n t i c i p a t i o n o f a n i n c r e a s e in s h o p p e r s . T h e F e d e r a l R e s e r v e c u t s i n t e r e s t r a t e s w h e n it
b e l i e v e s t h e e c o n o m y is a t risk f o r w e a k g r o w t h , a n d r a i s e s i n t e r e s t r a t e s w h e n it f e e l s t h a t i n f l a t i o n
is t h e g r e a t e r risk. By f i g u r i n g o u t t h e c h a n c e s o f v a r i o u s u n c e r t a i n e v e n t s , w e a r e b e t t e r p r e p a r e d
t o m a k e t h e m o r e d e s ir a b le c h o ic e s . T h is c h a p t e r p re s e n t s t h e e s s e n tia l p r o b a b i l i t y to o ls n e e d e d t o
f r a m e a n d a d d r e s s m a n y r e a l - w o r l d issues in v o l v i n g u n c e r t a i n t y . U n c e r t a i n t y d e s c r ib e s a s i tu a tio n
w h e r e a v a r i e t y o f e v e n t s a r e p o s s ib le . P r o b a b i l i t i e s tell us h o w o f t e n w e c a n a n t i c i p a t e o b s e r v i n g
f o u n d a t i o n f o r s t a t i s t i c a l i n f e r e n c e , a n d n u m e r o u s c o n c e p t s i n t r o d u c e d in t h i s c h a p t e r a r e e s s e n t i a l
fo r u n d e r s ta n d in g later c h a p te rs .
I N T R O D U C T O R Y CASE
Sportswear Brands
A nnabel Gonzalez is ch ie f retail analyst at L on g m e a d o w Consultants, a m a rketin g firm . One aspect o f her jo b is to
track sports-apparel sales and uncover any particu la r trends th a t may be u n fo ld in g in th e industry. Recently, she
has been fo llo w in g U nder Arm our, Inc., th e p io n e e r in th e com pression-gear m arket. Com pression garm ents are
m ea nt to keep m o isture away fro m a wearer's b o d y d u rin g a th le tic activities in w arm and cool weather. Under
A rm o u r has experienced expone ntial g ro w th since th e firm w e n t p u b lic in N ovem ber 2005. However, Nike, Inc.,
and Adidas Group, w ith 18% and 10% m arket shares, respectively, have aggressively entered th e com pression-
gear m arket ( T h e W a ll S tr e e t J o u r n a l, O ctober 2 3 , 2007).
As p a rt o f her analysis, A nnabel w o u ld first like to exam ine w h e th e r th e age o f th e cu sto m e r m atters w hen
bu yin g com pression cloth in g . Her initial feeling is th a t th e U nder A rm o u r brand attracts a y o u n g e r customer,
whereas th e m ore established companies, Nike and Adidas, dra w an o ld e r clientele. She believes th is in fo rm a tio n
is relevant to advertisers and retailers in th e sp o rtin g -g o o d s in d u s try as w ell as to som e in th e financial c o m m u
nity. She collects data on 600 recent purchases in th e com pression-gear m arket. She cross-classifies th e data by
age g ro u p and brand name, as show n in Table 4.1.
TABLE 4.1 Purchases of Compression Garments Based on Age and Brand Name
B rand N a m e
In order to define an event and assign the appropriate probability to it, it is useful to first
establish some terminology and impose some structure on the situation.
An experiment is a trial that results in any one of several possible outcomes. The di
versity of the outcomes of an experiment is due to the uncertainty of the real world. When
you purchase a new computer, there is no guarantee as to how long it will last before any
repair work is needed. It may need repair in the first year, in the second year, or after two
years. You can think of this as an experiment because the actual outcome will be deter
mined only over time. Other examples of an experiment include whether a roll of a fair
die will result in a value of 1, 2, 3, 4, 5, or 6; whether the toss of a coin results in heads
or tails; whether a project is finished early, on time, or late; whether the economy will
improve, stay the same, or deteriorate; whether a ball game will end in a win, loss, or tie.
A sample space, denoted by S, of an experiment records all possible outcomes of
the experiment. For example, suppose the sample space representing the letter grade in
a course is given by S = {A , B, C, D, F}. If the teacher also gives out an I (incomplete)
grade, then S is not valid because all outcomes of the experiment are not included in S.
The sample space for an experiment need not be unique. For example, in the above ex
periment, we can also define the sample space with just P (pass) and F (fail) outcomes, or
S = {P, F}.
EXAMPLE 4.1
A snowboarder competing in the Winter Olympic Games is trying to assess her
probability of winning a medal in her event, the ladies’ halfpipe. Construct the
appropriate sample space.
Events
An event is a subset of the sample space. A simple event consists of just one of the pos
sible outcomes of an experiment. Getting an A in a course is an example of a simple
event. An event may also be a subset of outcomes of an experiment. For example, we can
define an event as getting a passing grade in a course; this event is formed by the subset
of outcomes, A, B, C, and D.
An event is a subset of the sample space. It can be a simple event consisting of one
outcome or it can be a subset of several outcomes.
Let us define two events from Example 4.1, where one event represents “earning a medal”
and the other denotes “failing to medal.” These events are exhaustive because they in
clude all outcomes in the sample space. In other words, they exhaust the entire sample
space. This contrasts with the earlier grade-distribution example, where the events of get
ting grades A and B are not exhaustive because they do not include many feasible grades
in the sample space. However, the events P and F, defined as pass and fail, respectively,
are exhaustive.
Another important probability concept concerns m utually exclusive events. Suppose
we define the two events “at least earning a silver medal” (outcomes of gold and silver)
and “at most earning a silver medal” (outcomes of silver, bronze, no medal). These two
events are exhaustive because no outcome of the random experiment is omitted. However,
in this case, the events are not mutually exclusive because the outcome “silver” appears in
both events. For two mutually exclusive events, the occurrence of one event precludes the
occurrence of the other. Going back to the grade-distribution example, while the events of
getting grades A and B are not exhaustive, they are mutually exclusive, since you cannot
possibly get an A as well as a B in the same course. However, getting grades P and F are
exhaustive and mutually exclusive. Similarly, the events defined as “at least earning a sil
ver medal” and “at most earning a bronze medal” are exhaustive and mutually exclusive.
Events are exhaustive if all possible outcomes of a random experiment are included
in the events.
Events are m utually exclusive if they do not share any common outcome of a ran
dom experiment.
Given a sample space consisting of simple events, we can define events and then combine
events to form new events. The union of two events, denoted A ∪ B, is the event consist
ing of all outcomes in A or B. A useful way to illustrate these concepts is through the use
of a Venn diagram, named after the British mathematician John Venn (1834– 1923). Fig
ure 4.1 shows a Venn diagram where the rectangle represents the sample space S and the
two circles represent events A and B. The union A ∪ B is the portion in the Venn diagram
that is included in either A or B.
The intersection of two events, denoted A ∩ B, is the event consisting of all outcomes in
A and B. Figure 4.2 depicts the intersection of two events A and B. The intersection A ∩ B
is the portion in the Venn diagram that is included in both A and B.
FIGURE 4 .2 The intersection of tw o events, A ∩ B
The complement of event A, denoted A c, is the event consisting of all outcomes in the
sample space S that are not in A. In Figure 4.3, Ac is everything in S that is not included in A.
C O M B I N I N G EVENTS
• The union of two events, denoted A ∪ B, is the event consisting of all outcomes
in A or B.
• The intersection of two events, denoted A ∩ B, is the event consisting of all
outcomes in A and B.
• The complement of event A, denoted A c, is the event consisting of all outcomes
in the sample space S that are not in A.
EXAMPLE 4 .2
Recall that the snowboarder’s sample space from Example 4.1 is defined as S =
{gold, silver, bronze, no m edal}. Now suppose the snowboarder defines the follow
ing three events:
• A = {gold, silver, bronze} or event A denotes earning a medal;
• B = {silver, bronze, no medal} or event B denotes earning at most a silver
medal; and
• C = {no medal} or event C denotes failing to medal.
a. Find A ⋃ B and B ⋃ C.
b. Find A ∩ B and A ∩ C.
c. Find Bc.
SO LU TIO N :
a. The union of A and B denotes all outcomes common to A or B ; here, the event
A ⋃ B = {gold, silver, bronze, no medal}. Note that there is no double count
ing of the outcomes “silver” or “bronze” in A ⋃B. Similarly, we have the event
B ⋃ C = {silver, bronze, no m edal}.
b. The intersection of A and B denotes all outcomes common to A and B: here,
the event A ∩ B = {silver, bronze}. The event A ∩ C = Ø , where Ø denotes
the null (empty) set; no common outcomes appear in both A and C.
c. The complement of B denotes all outcomes in S that are not in B ; here, the
event Bc = {gold}.
Reading from the table we can readily see, for instance, that the athlete assesses that there
is a 15% chance that she will earn a silver medal, or P({silver}) = 0.15. We should note
that all the probabilities are between the values of zero and one, and when we sum the
probabilities of these simple events we obtain the value one. This example demonstrates
the two defining properties of probability.
T H E T W O D E F I N I N G P R O P E R T I E S OF A P R O B A B I L I T Y
1. The probability of any event A is a value between 0 and 1, 0 ≤ P(A ) ≤ 1.
2. The sum of the probabilities of any list of mutually exclusive and exhaustive
events equals 1.
Suppose the snowboarder wants to calculate the probability of earning a medal. In Exercise 4.2
we defined “earning a medal” as event A, so the probability statement takes the form P(A ). We
calculate this probability by summing the probabilities of the outcomes in A, or equivalently,
SO LU TIO N:
a. The probability that event B or event C occurs is
P(B ∪ C) = P
{silver}) + P ((bronze}) + P({no medal})
= 0.15 + 0.20 + 0.55 = 0.90.
EXAMPLE 4 .4
The frequency distribution in Table 4.3 summarizes the ages of the richest
400 Americans. Suppose we randomly select one of these individuals.
a. What is the probability that the individual is between 50 to 60 years old?
b. What is the probability that the individual is younger than 60 years old?
c. What is the probability that the individual is at least 80 years old?
Ages Frequency
30 up to 40 7
40 up to 50 47
50 up to 60 90
60 up to 70 109
70 up to 80 93
80 up to 90 45
90 up to 100 9
EXAMPLE 4.5
Suppose our random experiment consists of rolling a six-sided die. Then we can
define the appropriate sample space as S = {1, 2, 3, 4, 5, 6}.
a. What is the probability that we roll a 2?
b. What is the probability that we roll a 2 or 5?
c. What is the probability that we roll an even number?
C O N V E R T IN G AN O D D S RATIO TO A P RO BA BILITY
Thus, with odds for winning the Super Bowl of 1:24, we can solve for the probability
of the Steelers’ winning as: 1/(1 + 24) = 1/25 or 0.04. Moreover, the bet’s anticipated
profit is $0 because (0.04 probability of winning) × ($24 profit if the wager is won) +
(0.96 probability of losing) × ( − $1 if the wager is lost) = 0.96 + ( − 0.96) = 0.
This is an example of an expected value calculation, which we discuss further in
Chapter 5. We would also like to point out that sports betting odds are usually displayed in
various formats, including American, British, or European formats; the details are beyond
the scope of this chapter.
EXAMPLE 4.6
Days prior to the 2009 Super Bowl, the Pittsburgh Steelers’ odds for beating the
Arizona Cardinals increased to approximately 2:1. What was the probability of the
Steelers’ winning just prior to the Super Bowl?
SOLUTION: The probability that the Steelers would win the Super Bowl rose to
(Note: The Steelers did win the Super Bowl, but just barely, scoring the winning
touchdown with 35 seconds left in the game.)
Similarly, we can convert a probability to an odds ratio using the following generalization:
C O N V E R T I N G A P R O B A B IL IT Y TO AN O D D S RATIO
If P(A ) denotes the probability of an event A occurring, and P(A ) does not equal
zero or one, then:
The odds fo r A occurring equal and
EXAMPLE 4.7
The summer of 2008 proved to be another difficult period for travelers. New York’s
Kennedy Airport topped the list with the lowest on-time arrival rate: the likelihood
that a plane arrived on-time occurred only 56% of the time (The Wall Street Journal,
September 9, 2008). Travelers at Atlanta’s Airport fared a bit better, where the on-
time arrival rate was 74%.
a. Calculate the odds for a plane arriving on-time at New York’s Kennedy Airport.
b. Calculate the odds for a plane arriving on-time at Atlanta’s Airport.
SOLUTION:
a. First, given an on-time arrival probability of 0.56 for New York’s Kennedy
Airport we find
or, we would report the odds for arriving on-time as 1.27 to 1. Note that given
an odds for arriving on-time as 1.27:1, we can deduce P({on-time}) as
b. We calculate the odds for on-time arrival at Atlanta’s Airport given a probabil
ity of 0.74 as
EXERCISES 4.1
5. A sample space, S, yields four simple events, A , B, C, and D, Massachusetts, 15,164 patients were asked to respond to
questions designed to detect depression (The B o s to n G lobe,
such tha t P(A ) = 0.35, P(B) = 0.10, and P (C ) = 0.25.
June 11, 2008). The survey produced the follow ing results.
a. Find P(D).
Diagnosis Number
b. Find P(Cc).
c. Find P(A ∪ B). Mild 3,257
Moderate 1,546
Applications Moderately Severe 975
6. You apply for a position at tw o firms. Let event A represent Severe 773
the outcome o f getting an offer from the first firm and No Depression 8,613
event B represent the outcom e o f getting an offer from the
a. What is the probability th a t a randomly selected patient
second firm.
suffered from mild depression?
a. Explain w hy events A and B are not exhaustive.
b. W hat is the probability tha t a random ly selected patient
b. Explain why events A and B are not m utually exclusive.
did not suffer from depression?
7. An alarming number o f U.S. adults are either overweight
c. What is the probability tha t a randomly selected patient
or obese. The distinction between overweight and obese
suffered from moderately severe to severe depression?
is made on the basis o f body mass index (BMI), expressed
d. Given that the national figure for moderately severe to
as w eight/height2. An adult is considered overweight if the
severe depression is approximately 6.7%, does it appear
BMI is 25 or more but less than 30. An obese adult will have
that there is a higher rate o f depression in this summer
a BMI o f 30 or greater. According to the 2003–2004 National
resort community? Explain.
Health and Nutrition Examination Survey, 34.1% o f the adult
population in the U.S. is overweight and 32.2% is obese. Use 10. On Sunday, July 11, 2010, Spain and the Netherlands played
this information to answer the following questions. in the 2010 World Cup Final in Johannesburg. On the eve of
the final, many betting lines were offering Spain's odds for
a. W hat is the probability tha t a randomly selected adult is
w inning at 15:8 (O d d s c h e c k e r website).
either overweight or obese?
a. Spain won the World Cup. Suppose you had bet $1,000
b. W hat is the probability tha t a randomly selected adult is
on Spain. What was your net gain? If Spain had lost, w hat
neither overweight nor obese?
w ould have been your net loss?
c. Are the events "overweight" and "obese" exhaustive?
b. W hat was the im plied probability o f Spain w inning the
d. Are the events"overw eight"and "obese" m utually
final?
exclusive?
11. Prior to the Academy Awards ceremony in 2009, the United
8. Many com munities are finding it more and more difficult
Kingdom bookmaker Ladbrokes reported the follow ing odds
to fill municipal positions such as tow n administrators,
for w inning an Oscar in the category o f best actress (The W a ll
finance directors, and treasurers. The follow ing table shows
S tre e t J o u rn a l, February 2 0 , 2009).
the percentage o f municipal managers by age group in the
United States for the years 1971 and 2006. Best Actress Movie Odds
T H E C O M P L E M E N T RULE
T h e c o m p le m e n t r u le s t a t e s t h a t t h e p r o b a b i l i t y o f t h e c o m p l e m e n t o f a n e v e n t , P (A c) ,
i s e q u a l t o o n e m i n u s t h e p r o b a b i l i t y o f t h e e v e n t , o r e q u i v a l e n t l y , P (A c) = 1 − P (A ).
The complement rule is quite straightforward and rather simple, but it is widely used and
powerful.
EXAMPLE 4.8
In 2007, 45% of women ages 25 to 34 had a college degree, compared with 36% of
young men (The Wall Street Journal, May 29, 2008).
a. What is the probability that a randomly selected woman between the ages of
25 to 34 does not have a college degree?
b. What is the probability that a randomly selected man between the ages of 25 to
34 does not have a college degree?
SOLUTION:
a. Let’s define event A as the event that a randomly selected woman between
the ages of 25 to 34 has a college degree, thus P(A ) = 0.45. In this problem
we are interested in the complement of A or P(Ac). So P(A c) = 1 − P(A ) =
1 − 0.45 = 0.55.
b. Similarly, we define event B as the outcome that a randomly selected man
between the ages of 25 to 34 has a college degree, so P(B) = 0.36. Thus,
P (Bc) = 1 − P(B ) = 1 − 0.36 = 0.64.
P(A ∩ B). When implementing the addition rule, we sum P(A ) and P(B) and then subtract
P(A ∩∩ B) from this sum.
T H E A D D I T I O N RULE
The addition rule states that the probability that A or B occurs, or that at least one
of these events occurs, is equal to the probability that A occurs, plus the probability
that B occurs, minus the probability that both A and B occur, or equivalently,
P(A ∪ B ) = P(A ) + P(B) − P(A ∩ B).
EXAMPLE 4.9
Anthony feels that he has a 75% chance of getting an A in Statistics and a 55%
chance of getting an A in Managerial Economics. He also believes he has a
40% chance of getting an A in both classes.
a. What is the probability that he gets an A in at least one of these courses?
b. W hat is the probability that he does not get an A in either of these courses?
SO LU TIO N:
a. Let P(A S) correspond to the probability of getting an A in Statistics and P (AM)
correspond to the probability of getting an A in Managerial Economics.
Thus, P(A S) = 0.75 and P(A M) = 0.55. In addition, there is a 40% chance that
Anthony gets an A in both classes, or P (A S ∩ AM) = 0.40. In order to find the
probability that he receives an A in at least one of these courses, we calculate:
P(A s ∪A m) = P(A S) + P(A m) − P(AS ∩ AM) = 0.75 + 0.55 − 0.40 = 0.90.
b. The probability that he does not receive an A in either of these two courses is
actually the complement of the union of the two events or P ((A S ∩ A M )c).
We
calculated the union in part a, so using the complement rule we have
P((AS ∪ AM)c) = 1 − P(AS ∪ AM) = 1 − 0.90 = 0.10.
For mutually exclusive events A and B, the probability of their intersection is zero,
P(A ∩ B ) = 0. We need not concern ourselves with double-counting, and therefore, the
probability of the union is simply the sum of the two probabilities.
T H E A D D I T I O N RUL E F O R M U T U A L L Y E X C L U S I V E E V E N T S
If A and B are mutually exclusive events, then P (A ∩ B ) = 0 and, therefore, the
addition rale simplifies to P(A ∪ B ) = P(A ) + P(B ).
EXAMPLE 4.10
Samantha Greene, a college senior, contemplates her future immediately after gradu
ation. She thinks there is a 25% chance that she will join the Peace Corps and teach
English in Madagascar for the next 27 months. Alternatively, she believes there is a
35% chance that she will enroll in a full-time law school program in the United States.
a. What is the probability that she joins the Peace Corps or enrolls in law school?
b. What is the probability that she does not choose either of these options?
SOLUTION:
a. We can write the probability that Samantha joins the Peace Corps as P(A ) = 0.25
and the probability that she enrolls in law school as P(B ) = 0.35. Immediately after
college, Samantha cannot choose both of these options. This implies that these
events are mutually exclusive, so P(A ∩ B ) = 0. Thus, when solving for the prob
ability that Samantha joins the Peace Corps or enrolls in law school, P(A ∪B), we
can simply sum P(A ) and P(B): P(A ∪ B) = P(A ) + P(B) = 0.25 + 0.35 = 0.60.
b. In order to find the probability that she does not choose either of these options,
we need to recognize that this probability is the complement of the union of
the two events, or P((A ∪ B )c) . Therefore, using the complement rule, we have
P((A ∪ B)c) = 1 − P(A ∪ B) = 1 − 0.60 = 0.40.
LO 4 . 4 Conditional Probability
Calculate and
In business applications, the probability of interest is often a conditional probability. Ex
interpret a
amples include the probability that the housing market will improve conditional on the
conditional
Federal Reserve taking remedial actions; the probability of making a six-figure salary
probability.
conditional on getting an MBA; the probability that a company’s stock price will go up
conditional on higher-than-expected profits; the probability that sales will improve condi
tional on the firm launching a new innovative product.
L et’s use an example to illustrate the concept of conditional probability. Suppose
the probability that a recent business college graduate finds a suitable job is 0.80.
The probability o f finding a suitable job is 0.90 if the recent business college gradu
ate has prior work experience. This type of probability is called a conditional p ro b
a b ility , where the probability o f an event is conditional on the occurrence of another
event. If A represents “finding a jo b ” and B represents “prior work experience,”
then P (A ) = 0.80 and the conditional probability is denoted as P (A | B ) = 0.90.
The vertical mark | means “given that” and the conditional probability is typically
read as “the probability of A given B .” In the above example, the probability of
finding a suitable job increases from 0.80 to 0.90 when conditioned on prior work
experience. In general, the conditional probability, P(A|B), is greater than the
unconditional (m a rg in a l) p ro b a b ility , P (A ), if B exerts a positive influence on A.
Similarly, P(A| B ) is less than P (A ) when B exerts a negative influence on A. Finally,
if B exerts no influence on A, then P (A| B ) equals P (A ).
As we will see later, it is important that we write the event that has already occurred
after the vertical mark, since in most instances P(A|B ) ≠ P (B |A ). In the above example
P (B |A ) would represent the probability of prior work experience conditional on having
found a job.
We again rely on the Venn diagram in Figure 4.6 to explain the conditional probability.
EXAMPLE 4 .11
Economic globalization is defined as the integration of national economies into the
international economy through trade, foreign direct investment, capital flows, mi
gration, and the spread of technology. Although globalization is generally viewed
favorably, it also increases the vulnerability of a country to economic conditions of
the other country. An economist predicts a 60% chance that country A will perform
poorly and a 25% chance that country B will perform poorly. There is also a 16%
chance that both countries will perform poorly.
a. What is the probability that country A performs poorly given that country B
performs poorly?
b. What is the probability that country B performs poorly given that country A
performs poorly?
c. Interpret your findings.
a.
b.
I N D E P E N D E N T VERSUS D E P E N D E N T EVENTS
Two events, A and B , are independent if and only if P(A \ B ) = P(A ) or, equiva
lently, P(B |A ) = P(B ). Otherwise, the events are dependent.
EXAMPLE 4 .1 2
Suppose that for a given year there is a 2% chance that your desktop computer will
crash and a 6% chance that your laptop computer will crash. Further, there is a
0.12% chance that both computers will crash. Is the reliability of the two computers
independent of each other?
S O L U T IO N : Let event D represent the outcome that your desktop crashes and event L
represent the outcome that your laptop crashes. Therefore, P(D ) = 0.02, P(L) = 0.06,
and P(D ∩ L) = 0.0012. The reliability of the two computers is independent because
In other words, if your laptop crashes, it does not alter the probability that your
desktop also crashes. Equivalently,
= P ( B | A ) P (A ). S i n c e w e c a l c u l a t e t h e p r o d u c t o f t w o p r o b a b i l i t i e s t o f i n d t h e j o i n t p r o b
a b i l i t y , t h e r e s u l t i n g e q u a t i o n s a r e c a l l e d t h e m u lt ip lic a t io n r u l e f o r p r o b a b i l i t i e s .
T H E M U L T I P L I C A T I O N R UL E
The m ultiplication rule states that the probability that A and B both occur, a jo in t
probability, is equal to the probability that A occurs given that B has occurred times
the probability that B occurs, that is, P(A ∩ B ) = P(A| B )P(B). Equivalently, we can
also arrive at the joint probability as P(A ∩ B) = P(B |A )P(A ).
EXAMPLE 4.1 3
A stockbroker knows from past experience that the probability that a client owns
stocks is 0.60 and the probability that a client owns bonds is 0.50. The probability
that the client owns bonds if he/she already owns stocks is 0.55.
a. What is the probability that the client owns both of these securities?
b. Given that the client owns bonds, what is the probability that the client owns stocks?
SOLUTION:
a. Let event A denote the outcome that a client owns stocks and event B as the out
come that a client owns bonds. Thus, the unconditional probabilities that the cli
ent owns stocks and that the client owns bonds are P(A ) = 0.60 and P(B) = 0.50,
respectively. The conditional probability that the client owns bonds given that
he/she owns stocks is P(B |A ) = 0.55. We calculate the probability that the client
owns both of these securities as P(A ∩ B) = P(B |A )P(A ) = 0.55 × 0.60 = 0.33.
b. We need to calculate the conditional probability that the client owns stocks given
that he/she owns bonds, or P(A |B). Using the formula for conditional probability
and the answer from part (a), we find
T H E M U L T I P L I C A T I O N RUL E F O R I N D E P E N D E N T E V E N T S
The multiplication rule for independent events dictates that the joint probabil
ity of A and B equals the product of the individual probabilities of A and B, or
P(A ∩ B ) = P(A )P (B).
EXAMPLE 4 .14
The probability of passing the Level 1 CFA (Chartered Financial Analyst) exam is
0.50 for John Campbell and 0.80 for Linda Lee. The prospect of John’s passing the
exam is completely unrelated to Linda’s success on the exam.
a. What is the probability that both John and Linda pass the exam?
b. What is the probability that at least one of them passes the exam?
SOLUTION:
We can write the unconditional probabilities that John passes the exam and that
Linda passes the exam as P(J ) = 0.50 and P(L) = 0.80, respectively.
a. Since we are told that John’s chances of passing the exam are not influenced by
Linda’s success at the exam, we can conclude that these events are independent,
so P(J ) = P(J |L) = 0.50 and P(L) = P(L |J ) = 0.80. Thus, when solving for the
probability that both John and Linda pass the exam, we calculate the product of
the unconditional probabilities or P(J ∩ L) = P(J) × P(L) = 0.50 × 0.80 = 0.40.
b. We calculate the probability that at least one of them passes the exam as:
P(J ∪L) = P(J ) + P(L ) − P(J ∩ L ) = 0.50 + 0.80 − 0.40 = 0.90.
EXERCISES 4.2
Concepts a. What is the probability that at least one o f the stocks w ill
rise in price?
12. Let P(A ) = 0.65, P (B) = 0.30, and P(A | B) = 0.45.
b. Are events A and B m utually exclusive? Explain.
a. Calculate P (A ∩ B).
c. Are events A and B independent? Explain.
b. Calculate P ( A ∪ B ).
22. Despite governm ent bailouts and stimulus money,
c. Calculate P (A |B).
unem ploym ent in the U.S. had not decreased significantly
13. Let P (A ) = 0.55, P (B ) = 0.30, and P (A ∩ B). = 0.10.
as economists had expected (US N e w s a n d W o r ld R ep o rt, July
a. Calculate P ( A |B). 2 , 2010). Many analysts predicted only an 18% chance o f
b. Calculate P ( A ∪ B). a reduction in unem ploym ent. However, if Europe slipped
c. Calculate P((A ∪ B)c). back into a recession, the probability o f a reduction in U.S.
14. Let A and 6 be m utually exclusive w ith P(A) = 0.25 and unem ploym ent would drop to 0.06.
c. What is the probability th a t neither A nor B takes place? b. W hat is the probability tha t a student w ill neither do
hom ework regularly nor w ill pass the course?
17. Let P(A) = 0.15, P(B) = 0.10, and P ( A ∩ B) = 0.05.
c. Are the events "pass the course" and "do hom ework
a. Are A and B independent events? Explain.
regularly" m utually exclusive? Explain.
b. Are A and B m utually exclusive events? Explain.
d. Are the events "pass the course" and "do hom ework
c. What is the probability th a t neither A nor B takes place?
regularly" independent? Explain.
18. Consider the follow ing probabilities: P(A) = 0.25,
24. Records show that 5% o f all college students are foreign
P(Bc) = 0.40, and P ( A ∩ B ) = 0.08. Find:
students w ho also smoke. It is also known th a t 50% o f
a. P(B )
all foreign college students smoke. W hat percent o f the
b. P (A |B) students at this university are foreign?
c. P (B|A)
25. An analyst estimates th a t the probability o f default on a
19. Consider the follow ing probabilities: P(Ac) = 0.30, seven-year AA rated bond is 0.06, w hile tha t on a seven-year
P (B ) = 0.60, and P (A ∩ Bc) = 0.24. Find: A rated bond is 0.13. The probability th a t they w ill both
a. P(A|Bc) default is 0.04.
b. P(Bc|A) a. W hat is the probability th a t at least one o f the bonds
c. Are A and B independent events? Explain. defaults?
20. Consider the follow ing probabilities: P (A ) = 0.40, b. W hat is the probability th a t neither the seven-year AA
P (B ) = 0.50, and P ( Ac ∩ Bc)= 0.24. Find: rated bond nor the seven-year A rated bond defaults?
a. P(Ac|Bc)
b. P(Ac∪ c. Given tha t the seven-year AA rated bond defaults, w hat
is the probability th a t the seven-year A rated bond also
Bc)
defaults?
c. P(A ∪ B )
26. In general, shopping online is supposed to be more
convenient than going to stores. However, according
Applications to a recent Harris Interactive poll, 87% o f people have
21. The probabilities th a t stock A w ill rise in price is 0.40 and experienced problems w ith an online transaction ( The W a ll
tha t stock B w ill rise in price is 0.60. Further, if stock B rises in Stre e t J o u r n a l, October 2 , 2007). Forty-two percent o f people
price, the probability th a t stock A w ill also rise in price is 0.80. who experienced a problem abandoned the transaction or
switched to a competitor's website. Fifty-three percent of 29. Since the fall o f 2008, m illions o f Americans have lost jobs
people who experienced problems contacted customer- due to the econom ic m eltdow n. A recent study shows
service representatives. th a t un em p loym e nt has not im pacted w h ite collar and
a. What percentage o f people did not experience problems blue collar workers equally (N e w s w e e k , April 2 0 , 2009).
w ith an online transaction? According to the Bureau o f Labor Statistics report, w h ile
b. What percentage o f people experienced problems w ith the national u n em p loym e nt rate is 8.5%, it is only 4.3%
an online transaction and abandoned the transaction or fo r those w ith a college degree. It is fair to assume tha t
switched to a competitor's website? 27% o f people in the labor force are college educated. You
have ju s t heard th a t another w orker in a large firm has
c. What percentage o f people experienced problems w ith
been laid off. W hat is the p ro b a b ility th a t the w orker is
an online transaction and contacted customer-service
college educated?
representatives?
30. A recent study challenges the media narrative that
27. Christine Wong has asked Dave and Mike to help her move
foreclosures are dangerously widespread (N e w York
into a new apartm ent on Sunday morning. She has asked
Times, March 2 , 2009). According to this study, 62% of all
them both in case one of them does not show up. From past
foreclosures were centered in only four states, namely,
experience, Christine knows that there is a 40% chance that
Arizona, California, Florida, and Nevada. The national average
Dave will not show up and a 30% chance that Mike will not
rate of foreclosures in 2008 was 0.79%. What percent of the
show up. Dave and Mike do not know each other and their
homes in the United States were foreclosed in 2008 and also
decisions can be assumed to be independent.
centered in Arizona, California, Florida or Nevada?
a. What is the probability tha t both Dave and Mike will
show up? 31. According to results from the Spine Patient Outcomes
Research Trial, or SPORT, surgery for a painful, common
b. What is the probability that at least one of them will
back condition resulted in significantly reduced back pain
show up?
and better physical function than treatm ent w ith drugs and
c. What is the probability that neither Dave nor Mike will
physical therapy (The W a ll S tre e t J o u rn a l, February 2 1 , 2008).
show up?
SPORT followed 803 patients, o f whom 398 ended up getting
28. According to a recent survey by tw o United Nations agencies surgery. After tw o years, o f those who had surgery, 63% said
and a nongovernmental organization, tw o in every three they had a major im provement in their condition, compared
women in the Indian capital o f New Delhi are likely to face w ith 29% among those w ho received nonsurgical treatment.
some form o f sexual harassment in a year (BBC W o rld N ew s,
a. What is the probability that a patient had surgery? What
July 9 , 2010). The study also reports tha t wom en who use
is the probability that a patient did not have surgery?
public transportation are especially vulnerable. Suppose the
b. What is the probability tha t a patient had surgery
corresponding probability of harassment for wom en who
and experienced a major im provement in his or her
use public transportation is 0.82. It is also known tha t 28% of
condition?
women use public transportation.
c. What is the probability that a patient received
a. What is the probability that a woman takes public
nonsurgical treatm ent and experienced a major
transportation and also faces sexual harassment?
improvem ent in his or her condition?
b. If a wom an is sexually harassed, w hat is the probability
that she had taken public transportation?
Each cell in Table 4.4 represents a frequency; for example, 174 customers under the age
of 35 purchased an Under Armour product, whereas 54 customers at least 35 years old
purchased an Under Armour product. Recall that we estimate an empirical probability
by calculating the relative frequency of the occurrence of the event. To make calculating
these probabilities less cumbersome, it is often useful to denote each outcome with letter
notation and calculate totals for each column and row as follows:
Thus, events A and Ac represent the outcomes “under 35 years” and “Under Armour,”
“Nike,” and “Adidas” “35 years and older,” respectively; events B 1, B2, and B3 stand for
the outcomes, respectively. In addition, after calculating row totals, it is now easier to
recognize that 396 of the customers were under 35 years old and 204 of the customers
were at least 35 years old. Similarly, column totals indicate that 228 customers purchased
Under Armour, 204 purchased Nike, and 168 purchased Adidas. Finally, the frequency
corresponding to the cell in the last column and the last row is 600. This value represents
the sample size, or the possible outcomes in the sample space. We arrive at this value by
either summing the values in the last column (396 + 204) or summing the values in the
last row (228 + 204 + 168).
The following example illustrates how to calculate probabilities when the data are
presented in the form of a contingency table.
EXAMPLE 4.15
Using the data in Table 4.4a, answer the following questions.
a. What is the probability that a randomly selected customer is younger than
35 years old?
b. What is the probability that a randomly selected customer purchases an Under
Armour garment?
c. W hat is the probability that a customer is younger than 35 years old and pur
chases an Under Armour garment?
d. What is the probability that a customer is either younger than 35 years old or
purchases an Under Armour garment?
e. W hat is the probability that a customer is under 35 years of age, given that the
customer made an Under Armour purchase?
SO LU TIO N :
a. there is a 66% chance that a
randomly selected customer is less than 35 years old.
c. there is a 29%
chance that a randomly selected customer is younger than 35 years old and
purchases an Under Armour garment.
d.
there is a 75% chance that a randomly selected customer is either younger than
35 years old or purchases an Under Armour garment. Alternatively, we can use
the addition rule to solve this problem as P(A ∪ B1) = P(A ) + P(B1) − P(A ∩ B1) =
0.66 + 0.38 − 0.29 = 0.75.
e. We wish to calculate the conditional probability, P(A| B 1). When the data are in
the form of a contingency table, calculating a conditional probability is rather
straightforward. We are given the information that the customer has already
purchased an Under Armour product, so the relevant number of outcomes
shrinks from 600 to 228. We can ignore all customers that made Nike or Adi
das purchases, or all outcomes in events B2 and B 3. Thus, of the 228 customers
who made an Under Armour purchase, 174 of them are under 35 years of age.
Therefore, the probability that a customer is under 35 years of age given that
the customer makes an Under Armour purchase is calculated as
= 0.76. Alternatively, we can use the conditional probability formula to solve
the problem as
T A B L E 4 .4 b C o n v e r t in g a C o n t i n g e n c y T a b le t o a J o in t P r o b a b ilit y T a b le
B rand N a m e
All the probabilities in the interior of the table represent joint probabilities. For instance,
the probability that a randomly selected person is under 35 years of age and makes an
Under Armour purchase, denoted P(A ∩ B1), is 0.29. Similarly, we can readily read from
this table that 12% of the customers purchase a Nike garment and are at least 35 years
old, o r P (A c ∩ B2) = 0.12.
The probabilities on the periphery of Table 4.4b represent unconditional probabilities. For
example, the probability that a randomly selected customer is under 35 years of age, P(A ) is
simply 0.66. Also, the probability of purchasing a Nike garment, P(B2), is 0.34.
Note that the conditional probability is basically the ratio of a joint probability to an
unconditional probability. Since the numerator is a joint probability,
P(A ∩ B 1), and the denominator is an unconditional probability, P(B 1). Let’s refer back to
the probability that we calculated earlier; that is, the probability that a customer is under
35 years of age, given that the customer already purchased an Under Armour product. As
shown earlier, the conditional probability is easily computed as
EXAMPLE 4.16
Given the data in Table 4.4b, what is the probability that a customer purchases an
Under Armour product, given that the customer is under 35 years of age?
EXAMPLE 4.17
Determine whether the age of a customer is independent of the brand name of the
product purchased.
S Y N O P S I S OF I N T R O D U C T O R Y CAS E
EXERCISES 4.3
33. Consider the follow ing jo in t probability table. f. Is jo b category independent o f w hether or n o t a worker
slept on th e job? Explain using probabilities.
B1 B2 B3 B4
35. A recent poll asked 16- to 21 -year-olds w h eth er or no t they
A 9% 22% 15% 20%
are likely to serve in the U.S. military. The fo llo w in g table,
Ac 3% 10% 9% 12% cross-classified by gender and race, reports the percentage
o f those polled w h o responded th a t they are likely or very
a. What is the probability that A occurs?
likely to serve in the active-duty military.
b. What is the probability that B2 occurs?
c. What is the probability that A c and B4 occur? Race
o f governm ent workers adm itted to falling asleep on d. Given th a t a respondent is w hite, w h a t is the probability
the job. Consider the follow ing contingency table that is th a t the respondent is male?
representative o f the survey results. e. Is gender independent o f race? Explain using probabilities.
36. Merck & Co. conducted a study to test the promise o f its a. W hat is the p robability th a t a random ly selected
experim ental AIDS vaccine ( T h e B o s to n G lo b e , September 22, household uses heating oil?
2007). Volunteers in th e study were all free o f the human b. W hat is th e p robability th a t a random ly selected
im m unodeficiency virus (HIV), which causes AIDS, at the household is delinq uent in paying its u tility bill?
start o f th e study, bu t all were at high risk for g e tting the c. W hat is th e p robability th a t a random ly selected
virus. Volunteers were either given th e vaccine or a dum m y household uses heating oil and is delin q u e n t in paying
shot: 24 o f 741 volunteers w h o g o t th e vaccine became its u tility bill?
infected w ith HIV, whereas 21 o f 762 volunteers w h o go t the
d. Given th a t a household uses heating oil, w h a t is the
du m m y shot became infected w ith HIV. The fo llo w in g table
probability th a t it is d e lin q u e n t in paying its u tility bill?
summarizes th e results o f the study.
e. Given th a t a household is delin q u e n t in paying its u tility
Vaccinated Dummy Shot bill, w h a t is the p robability th a t th e household uses
Infected 24 21 electricity?
sample o f 500 households was taken to investigate if the c. W hat is th e p robability th a t a random ly selected
type o f heating influences w h e th e r or n o t a household European wom an does n o t like the perfume?
is d e lin q u e n t in paying its u tility bill. The fo llo w in g table d. W hat is th e pro b a b ility th a t a random ly selected
reports th e results. American man does no t like th e perfume?
e. Are there gender differences in th e preference fo r the
Type of Heating
perfum e in (i) America, (ii) Europe, (iii) Asia? Explain
Delinquent Natural Heating
using probabilities.
in Payment? Gas Electricity Oil Propane
50 20 f. Are there international differences in the preference
Yes 15 10
fo r the perfum e for men and women? Explain using
No 240 130 20 15
probabilities.
Oftentimes the joint probabilities needed to compute the total probability are not explic
itly specified. Therefore, we use the multiplication rule to derive these probabilities from
the conditional probabilities as P(A ∩ B) = P(A| B )P (B ) and P(A ∩ Bc) = P(A|Bc) P(Bc).
T H E T O T A L P R O B A B I L I T Y R UL E C O N D I T I O N A L O N T W O O U T C O M E S
The total probability rule expresses the unconditional probability of an event,
P(A ), in terms of probabilities conditional on various mutually exclusive and ex
haustive events. The total probability rule conditional on two events, B and Bc, is
P (A ) = P ( A | B ) + P( A | B c),
or equivalently,
P (A ) = P(A | B )P(B) + P(A| Bc)P(Bc).
An intuitive way to express the total probability rule is with the help of a probability tree.
Whenever a random experiment can be broken down into stages, with a different aspect
of the result observed at each stage, we can use a probability tree to represent the various
possible sequences of observations. Both the addition and multiplication rules can be il
lustrated with a probability tree. Later, we will use an alternative method to systematically
solve for an unconditional probability based on the total probability rule with the help of
a probability table. The following example illustrates the mechanics of a probability tree
and table.
EXAMPLE 4 .1 8
Even though a certain statistics professor does not require attendance as part of a
student’s overall grade, she has noticed that those who regularly attend class have
a higher tendency to get a final grade of A. The professor calculates that there is an
80% chance that a student attends class regularly. Moreover, given that a student
attends class regularly, there is a 35% chance that the student receives an A grade;
however, if a student does not attend class regularly, there is only a 5% chance of an
A grade. Use this information to answer the following questions.
a. What is the probability that a student does not attend class regularly?
b. What is the probability that a student attends class regularly and receives an A
grade?
c. What is the probability that a student does not attend class regularly and re
ceives an A grade?
d. What is the probability that a student receives an A grade?
SOLUTION: W e f i r s t d e n o t e e v e n t A a s t h e s i m p l e e v e n t c o r r e s p o n d i n g t o t h e o u t c o m e
that a student receives an A grade and event R as the simple event corresponding to
the outcome that a student attends class regularly. From the above information, we
then have the following probabilities: P(R ) = 0.80, P(A| R ) = 0.35, and P(A| Rc) =
0.05. Figure 4.8 shows a probability tree that consists of nodes (junctions) and
branches (lines) where the initial node O is called the origin. The branches emanating
from O represent the possible outcomes that may occur at the first stage. Thus, at
stage 1 we have outcomes R and Rc originating from O. These outcomes become the
nodes at the second stage. The sum of the probabilities coming from any particular
node is equal to one.
Ac ∩ R P( A c ∩ R ) = 0.52
A ∩ Rc ( A ∩ R c ) = 0.01
a. Using the complement rule, if we know that there is an 80% chance that a stu
dent attends class regularly, P (R ) = 0.80, then the probability that a student
does not attend class regularly is 0.20, or P(Rc) = 1 − P(R) = 1 − 0.80 = 0.20.
An alternative method uses the tabular representation of probabilities. Table 4.5, referred
to as a probability table, contains all relevant probabilities that are directly or indirectly
specified in Example 4.18.
TABLE 4.5 Probability Table for Class Attendance and Final Grade in Statistics
Attendance Probability Conditional Probability Joint Probability
P(R) = 0.80 P(A | R) = 0.35 P(A ∩ R ) = 0.28
P(Rc) = 0.20 P(A|Rc) = 0.05 P(A ∩ Rc) = 0.01
P(R) + P(Rc) = 1 P(A ) = 0.29
Bayes' Theorem
The total probability rule is also needed to derive Bayes’ theorem, developed by the
Reverend Thomas Bayes (1702– 1761). Bayes’ theorem is a procedure for updating
probabilities based on new information. The original probability is an unconditional
probability called a p rio r pro bability in the sense that it reflects only what we know
now before the arrival of any new information. On the basis of new information, we
update the prior probability to arrive at a conditional probability called a posterior
probability.
Suppose we know that 99% of the individuals who take a lie detector test tell the
truth. Therefore, the prior probability of telling the truth is 0.99. Suppose an individual
takes the lie detector test and the results indicate that the individual lied. Bayes’ theorem
updates a prior probability to compute a posterior probability, which in the above example
is essentially a conditional probability based on the information that the lie detector has
detected a lie.
Let P(B) denote the prior probability and P(B |A ) the posterior probability. Note that
the posterior probability is conditional on event A, representing new information. In a
sense, we can think of Bayes’ theorem as a special case of a conditional probability.
Recall the conditional probability formula from Section 4.2:
In some instances we may have to evaluate P(B |A ), but we do not have explicit
information on P (A ∩ B ) or P(A ). However, given information on P(B), P(A| B) and
P (A |B c), w e c a n u s e t h e t o t a l p r o b a b i l i t y r u l e a n d t h e m u l t i p l i c a t i o n r u l e t o s o l v e f o r P ( B | A )
a s fo llo w s :
BAYES'THEOREM
Given a set of prior probabilities for an event of interest, upon the arrival of new
information, the rule for updating the probability of the event is Bayes’ theorem.
Here P(B ) is the prior probability and P(B |A ) is the posterior probability:
or equivalently,
EXAMPLE 4.19
In a lie-detector test, an individual is asked to answer a series of questions, while
connected to a polygraph (lie detector). This instrument measures and records sev
eral physiological responses of the individual on the basis that false answers will
produce distinctive measurements. Assume that 99% of the individuals who go in
for a polygraph test tell the truth. These tests are considered to be 95% reliable. In
other words, there is a 95% chance that the test will detect a lie if an individual actu
ally lies. Let there also be a 0.5% chance that the test erroneously detects a lie even
when the individual is telling the truth. An individual has just taken a polygraph test
and the test has detected a lie. What is the probability that the individual was actu
ally telling the truth?
SOLUTION: First we define some events and their associated probabilities. Let D
denote the simple event corresponding to the outcome that the polygraph detects a
lie and T represent the simple event corresponding to the outcome that an individual
is telling the truth. We are given that P(T) = 0.99, implying that P(T c) = 1 − 0.99
= 0.01. In addition, we formulate P(D |T c) = 0.95 and P(D | T ) = 0.005. We need
to solve for P (T | D ) when we are not explicitly given P(D ∩ T) and P(D). We can
use Bayes’ theorem to calculate:
Although we can use this formula to solve the problem directly, it is often easier to
solve it systematically with the help of a probability tree or a probability table; we
will use the probability table shown in Table 4.6 for calculations.
The first column presents prior probabilities and the second column shows related
conditional probabilities. We first compute the denominator of Bayes’ theorem by
using the total probability rule, P(D) = P(D ∩ T ) + P(D ∩ T c). Joint probabilities
are calculated as products of conditional probabilities with their corresponding prior
probabilities. For instance, in Table 4.6, in order to obtain P(D ∩ T ), we multiply
P(D | T ) with P(T ), which yields P(D ∩ T ) = 0.005 × 0.99 = 0.00495. Similarly,
we find P(D ∩ T c) = 0.95 × 0.01 = 0.00950. Thus, according to the total probability
rule. P(D ) = 0.00495 + 0.00950 = 0.01445. Finally,
The prior probability of an individual telling the truth is 0.99.
However, given the new information that the polygraph detected the individual tell
ing a lie, the posterior probability of this individual telling the truth is now revised
downward to 0.34256.
So far we have used the total probability rule as well as Bayes’ theorem based on two
mutually exclusive and exhaustive events, namely, B and B c. We can easily extend the
analysis to include n mutually exclusive and exhaustive events, B 1, B2, . . . , Bn.
E X T E N S I O N S OF T H E T O T A L P R O B A B I L I T Y RULE
AND BAYES'THEOREM
Let a sample space be partitioned into n mutually exclusive and exhaustive events,
B 1, B2, . . . , Bn.
or equivalently,
EXAMPLE 4 .2 0
Scott Myers is a security analyst for a telecommunications firm called Webtalk. Al
though he is optimistic about the firm’s future, he is concerned that its stock price will
be hugely affected by the condition of credit flow in the economy. He believes that the
probability is 0.20 that credit flow will improve significantly, 0.50 that it will improve
only marginally, and 0.30 that it will not improve at all. He also estimates that the
probability that the stock price of Webtalk will go up is 0.90 with significant improve
ment in credit flow in the economy, 0.40 with marginal improvement in credit flow in
the economy, and 0.10 with no improvement in credit flow in the economy.
a. Based on Scott’s estimates, what is the probability that the stock price of
Webtalk goes up?
b. If we know that the stock price of Webtalk has gone up, what is the probability
that credit flow in the economy has improved significantly?
SOLUTION: As always, we first define the relevant events and their associated
probabilities. Let S, M, and N denote significant, marginal, and no improvement
in credit flow, respectively. Then P(S ) = 0.20, P(M ) = 0.50, and P(N) = 0.30. In
addition, allow G to denote an increase in stock price, we formulate P(G| S) = 0.90,
P(G | M ) = 0.40, and P(G | N ) = 0.10. We need to calculate P(G) in part (a) and
P(S|G
) in part (b). Table 4.7 aids in assigning probabilities.
EXERCISES 4.4
P(B) = 0.85 P(A| B) = 0.05 P(A ∩ B ) = P(B |A) = 44. An analyst expects th a t 20% o f all p u b lic ly trad e d com panies
w ill experience a de clin e in earnings n e xt year. The analyst
P(Bc) = P(A|Bc) = 0.80 P( A ∩ B c ) = P(Bc|A) =
has d e ve lop e d a ratio to h e lp forecast th is decline. If th e
Total = P(A ) = Total =
co m p a n y is headed fo r a decline, th e re is a 70% chance th a t
42. Let a sample space be partitioned into three m utually th is ratio w ill be negative. If th e co m p a n y is n o t headed fo r a
exclusive and exhaustive events, B 1, B 2, and B 3. C om plete decline, th e re is a 15% chance th a t th e ratio w ill be negative.
the following probability table. The analyst ra n d o m ly selects a co m p a n y and its ra tio is
negative. W h a t is th e p o sterio r p ro b a b ility th a t th e co m p a n y 47. An a n a ly s t th in k s th a t n e x t year th e re is a 20% chance th a t
w ill experience a decline? th e w o rld e c o n o m y w ill be g o o d , a 50% chance th a t it w ill
45. The State Police are try in g to crack d o w n on speeding on be n e u tra l, and a 30% chance th a t it w ill be p oor. She also
a p a rtic u la r p o rtio n o f th e Massachusetts Turnpike. To aid p re d ic ts p ro b a b ilitie s th a t a s ta rt-u p firm , C reative Ideas,
prom ises greater consistency and reliab ility. Specifically, states o f th e w o rld eco n om y. The fo llo w in g ta b le presents
4.5Counting Rules
In s e v e ra l a r e a s o f s ta tis tic s , in c lu d in g th e b in o m ia l d is tr ib u tio n d is c u s s e d in th e n e x t LO 4 .8
c h a p te r, th e c a lc u la tio n o f p r o b a b ilitie s in v o lv e s d e fin in g a n d c o u n tin g o u tc o m e s . H e re
U s e a c o u n t i n g rule
w e d is c u s s p r in c ip le s a n d s h o r tc u ts f o r c o u n tin g . S p e c ific a lly , w e e x p lo r e th e f a c to ria l,
t o s o l v e a p a rticu la r
c o m b in a tio n , a n d p e rm u ta tio n n o ta tio n s. c o u n tin g p ro b lem .
W h e n w e a r e in te r e s te d in c o u n tin g th e a r r a n g e m e n ts o f a g iv e n s e t o f n ite m s , w e
c a l c u l a t e n f a c t o r ia l, d e n o t e d n \ . I n o t h e r w o r d s , g i v e n n i t e m s , t h e r e a r e n ! w a y s o f a r
ra n g in g th e m . W e a p p ly th e fa c to ria l w h e n th e re a re n o g ro u p s — w e a re o n ly a rra n g in g a
g iv e n s e t o f n ite m s .
THE FACTORIAL F O R M U L A
T h e n u m b e r o f w a y s t o a s s ig n e v e r y m e m b e r o f a g r o u p o f s iz e n to n s lo ts is
c a lc u la te d u s in g th e f a c t o r i a l f o r m u l a :
n ! = n × (n − 1) × (n − 2 ) × (n − 3 ) × . . . × 1
B y d e f in itio n , 0 ! = 1.
EXAMPLE 4.21
A little-league coach has nine players on his team and he has to assign each of the
players to one of nine positions (pitcher, catcher, first base, etc.). In how many ways
can the assignments be made?
SOLUTION: The first player may be assigned to nine different positions. Then eight
positions remain. The second player can be assigned to eight different positions.
The third player can be assigned to seven different positions, and so on, until the
ninth and last player can be assigned in only one way. The total number of different
assignments is equal t o 9! = 9 × 8 × . . . × 1 = 362,880.
The com bination and perm utation form ulas apply to two groups of predetermined
size. We apply the combination formula when the order of the arrangement does not mat
ter, whereas we use the permutation formula when the order is important. Generally, we
look for a specific reference to “order” being important when employing the permutation
formula.
THE C O M B IN A T IO N F O R M U LA
The number of ways to choose x objects from a total of n objects, where the order in
which the x objects are listed does not matter, is calculated using the combination
form ula:
EXAMPLE 4.22
The little-league coach from Example 4.21 recruits three more players so that his
team has backups in case of injury. Now his team totals 12. In how many ways can
the coach select nine players from the 12-player roster?
THE P ER M U TA TIO N FO R M U LA
The number of ways to choose x objects from a total of n objects, where the order
in which the x objects is listed does matter, is calculated using the perm utation
form ula:
EXAMPLE 4.23
Now suppose the little league coach from Example 4.22 recognizes that the nine
positions of baseball are quite different. It matters whether one player is pitching
or whether that same player is in the outfield. The teammates that this player plays
with in any particular inning also matters. In how many ways can the coach assign
his 12-player roster to the nine different positions?
SO LU TIO N : This is a permutation problem because the order in which the coach
assigns the positions matters; that is, in one inning a player may catch, but in
another inning this same player may pitch. Even though the player is participating
in the two innings, the player’s position changes, thus it is a different arrangement.
We calculate the answer as follows:
Comparing the answers we obtained from Examples 4.22 and 4.23, we see there is a
big difference between the number of arrangements when the position of the player
does not matter versus the number of arrangements when the position is important.
EXERCISES 4.5
Concepts trial drug. In how many different ways can the researchers
select 10 patients to receive the trial drug from the total o f
48. Calculate the following values.
20?
a. 8! and 6!
52. There are 10 players on the local basketball team. The coach
b. 8C6
decides to randomly pick 5 players for the game.
C. 8P6
a. In how many different ways can the coach select 5
49. Calculate the following values. players to start the game if order does not matter?
a. 7! and 3! b. In how many different ways can the coach select 5
b. 7C3 players to start the game if order (the type o f position,
c. 8P6 i.e., po in t guard, center, etc.) matters?
53. David Barnes and his fiancee Valerie Shah are visiting Hawaii.
Applications At the Hawaiian Cultural Center in Honolulu, they are told
50. At a local elementary school, a principal is making random that 2 out o f a group o f 8 people w ill be randomly picked for
class assignments for her 8 teachers. Each teacher must a free lesson o f a Tahitian dance.
be assigned to exactly one job. In how many ways can the a. What is the probability tha t both David and Valerie get
assignments be made? picked for the Tahitian dance lesson?
51. Twenty cancer patients volunteer fo r a clinical trial. Ten o f b. What is the probability that Valerie gets picked before
the patients w ill receive a placebo and 10 w ill receive the David for the Tahitian dance lesson?
A University of Utah study examined 7,925 severely obese adults who had gastric bypass
surgery and an identical number of people who did not have the surgery (The Boston
Globe, August 23, 2007). The study wanted to investigate whether or not losing weight
through stomach surgery prolonged the lives of severely obese patients, thereby reducing
their deaths from heart disease, cancer, and diabetes.
Over the course of the study, 534 of the participants died. O f those who died, the
cause of death was classified as either a disease death (such as heart disease, cancer, and
diabetes) or a nondisease death (such as suicide or accident). Lawrence
Plummer, a research analyst, is handed Table 4.8 a contingency table
that summarizes the study’s findings:
Sample Numerous studies have documented the health risks posed to severely obese people—
those people who are at least 100 pounds overweight. Severely obese people, for instance,
Managerial typically suffer from high blood pressure and are more likely to develop diabetes. A Uni
versity of Utah study examined whether the manner in which a severely obese person lost
Report— weight influenced a person’s longevity. The study followed 7,925 patients who had stom
ach surgery and an identical number who did not have the surgery. Of particular interest
Linking Cause in this report are the 534 participants who died over the course of the study.
The deceased participants were cross-classified by the method in which they lost
of Death with weight and by the cause of their death. The possible outcomes for the method of losing
weight were either “no surgery” or “surgery,” and the possible outcomes for the cause of
the Method of death were either “disease death” (such as heart disease, cancer, and diabetes) or a “non-
disease death” (such as suicide or accident). Table 4. A shows the joint probability table.
Losing Weight
TABLE 4 .A Joint Probability Table of Deaths Cross-Classified by
Cause and Method o f Losing Weight
M e th o d o f Losing W e ig h t
The unconditional probabilities reveal that 0.60 of the deceased participants in the study
did not have surgery, while 0.40 of those who died had opted for the stomach surgery.
Of the 534 participants that died, the vast majority, 0.81, died from disease, whereas the
cause of death for the remainder was from a nondisease cause.
Joint probabilities reveal that the probability that a deceased participant had no surgery
and died from disease was 0.53; yet the probability that a deceased participant had sur
gery and died from disease was only 0.28. Using the unconditional probabilities and the
joint probabilities, it is possible to calculate conditional probabilities. For example, given
that a participant’s cause of death was from disease, the probability that the participant
did not have surgery was 0.65 (= 0.53/0.81). Similarly, of those participants who opted
for no surgery, the likelihood that their death was from disease was 0.88 (= 0.53/0.60).
A comparison of the conditional probabilities with the unconditional probabilities can
reveal whether or not the method of losing weight is independent of the cause of death. For
instance, there is an 81% chance that a randomly selected obese person dies from disease.
However, given that an obese person chooses to lose weight without surgery, the likelihood
that he/she dies from disease jum ps to 88%. Thus, this initial research appears to suggest
that a participant’s cause o f death is not independent o f his/her method o f losing weight.
Conceptual Review
LO 4 .1 Describe fundam ental probability concepts.
In order to assign the appropriate probability to an uncertain event, it is useful to estab
lish some terminology. An experim ent is a trial that results in one o f several possible
outcomes. A sample space, denoted S, o f an experiment contains all possible outcomes
of the experiment. A n event is any subset of outcomes o f an experiment, and is called a
simple event if it consists o f a single outcome. Events are considered exhaustive if all
possible outcom es o f an experim ent are included in the events. Events are considered
m utually exclusive if they do not share any com m on outcom e o f an experiment.
A probability is a num erical value that measures the likelihood that an uncertain
event occurs. It assum es a value betw een zero and one w here a value zero indicates an
im possible event and a value one indicates a definite event. The two defining proper
ties o f a probability are (1) the probability o f any event A is a value between 0 and 1,
0 ≤ P(A) ≤ 1, and (2) the sum o f the probabilities o f any list o f mutually exclusive and
exhaustive events equals 1.
We can extend the above rule where the sam ple space is partitioned into n m utu
ally exclusive and exhaustive events, B 1, B2, . . . , Bn. The total probability rule is:
P(A ) = P(A ∩ B 1 ) + P(A ∩ B2) + . . . . + P(A ∩ Bn), or equivalently, P(A ) = P(A | B l)P(B 1) +
P (A | B2)P (B 2) + • • • + P(A | Bn)P (Bn).
For the extended total probability rule, B ayes’ theorem, for any i = 1, 2 , . . . , n, is:
or
equivalently,
pugnacious teams (The Wall Street Journal, March 3 , 2009). 15,743 71,005 86,748
What are the odds for a fig h t occurring?
a. What is the probability tha t a randomly selected
57. Anthony Papantonis, owner o f Nauset Construction, is
patient experienced cardiac arrest during the
bidding on tw o projects, A and B. The probability that he
graveyard shift?
wins project A is 0.40 and the probability tha t he wins
project B is 0.25. Winning Project A and w inning Project B b. W h a t is t h e p r o b a b ility t h at a randomly selected patient
a. What is the probability tha t he wins project A or c. Given tha t a randomly selected patient experienced
project B? cardiac arrest during the graveyard shift, what is the
probability the patient survived for discharge?
b. W hat is the probability that he does not win either
project? d. Given tha t a randomly selected patient survived
for discharge, w hat is the probability the patient
58. Since the fall o f 2008, m illions o f Americans have lost jobs
experienced cardiac arrest during the graveyard shift?
due to the economic m eltdown. A recent study shows tha t
e. Is w h e th e r o r n o t a p a tie n t survives in d e p e n d e n t o f th e
unem ploym ent has not impacted males and females in
tim in g o f th e cardiac arrest? Explain using probabilities.
the same way (Newsweek, April 20,2009). According to a
Given your answer, what type of recommendations
Bureau of Labor Statistics report, 8.5% o f those w ho are
m ig ht you give to hospitals?
eligible to w ork are unemployed. The unem ploym ent rate
is 8.8% for eligible men and only 7.0% for eligible women. 6 1 . lt has been reported tha t women end up unhappier than
Suppose 52% of the eligible workforce in the U.S. consists men later in life, even though they start out happier (Yahoo
of men. News, August 1, 2008). Early in life, women are more likely
a. You have ju st heard that another worker in a large firm to fulfill their fam ily life and financial aspirations, leading
has b e e n la id off. W h a t is t h e p r o b a b ility t h a t th is w o rk e r to greater overall happiness. However, men report a higher
is a man? satisfaction w ith their financial situation and fam ily life,
and are thus happier than women in later life. Suppose
b. You have just heard tha t another worker in a large firm
the results o f the survey o f 300 men and 300 women are
has b e e n la id o ff. W h a t is t h e p r o b a b ility t h a t th is w o rk e r
presented in the follow ing table.
is a w o m a n ?
59. How much you smile in your younger days can predict your Response to the question "Are you satisfied w ith your financial
later success in marriage (msnbc.com, April 16, 2009). The and fam ily life?"
analysis is based on the success rate in marriage o f people Age
over age 65 and their smiles when they were only 10 years
Response by Women 20 to 35 35 to 50 Over 50
old. Researchers found that only 11% o f the biggest smilers 73 36
Yes 32
had been divorced, while 31% of the biggest frowners had
No 67 54 38
experienced a broken marriage.
Age a. What is the probability that both the U.S. economy and
the Asian countries will perform well?
Response by Men 20 to 35 35 to 50 Over 50 |
58 34 b. What is the unconditional probability tha t the Asian
Yes 38
countries will perform well?
No 92 46 32
c. What is the probability that the U.S. economy will
a. W h a t is th e p r o b a b ility t h a t a ra n d o m ly s e le c te d w o m a n perform well, given that the Asian countries perform
is satisfied w ith her financial and fam ily life? well?
b. W h a t is th e p ro b a b ility th a t a ra n d o m ly selected m an is 63. Apparently, depression significantly increases the risk of
satisfied w ith his financial and fam ily life? d e v e lo p in g d e m e n tia later in life (BBC News, July 6 , 2010). In
c. Does the above survey suggest that, for a woman, a recent study it was reported tha t 22% of those w ho had
satisfaction w ith life depends on age? Explain. depression w e n t on to d e ve lop d e m e n tia , com pared to o n ly
d. Does the above survey suggest that, for a man, 17% o f those w h o did n o t have depression. Suppose 10% o f
satisfaction w ith life depends on age? Explain. all people suffer from depression.
62. An analyst predicts that there is a 40% chance that the U.S. a. What is the probability o f a person developing
economy w ill perform well. If the U.S. economy performs dementia?
well, then there is an 80% chance tha t Asian countries w ill b. If a person has developed dementia, w hat is the
also perform well. On the other hand, if the U.S. economy probability that the person suffered from depression
performs poorly, the probability of Asian countries earlier in life?
perform ing well goes down to 30%.
CASE STUDIES
1. Find the likelihood that a w om an suffers from hypothyroidism during pregnancy and
later has a child with an I.Q. o f 85 or lower.
2. Determ ine the num ber o f children in a sample o f 100,000 that are likely to have an
I.Q. o f 85 or low er if the thyroid gland o f pregnant w omen is ignored.
3. Compare and com m ent on your answer to part b w ith the corresponding num ber if
all pregnant w omen are tested and treated for hypothyroidism.
LEARNING OBJECTIVES
After r e a d in g
th is c h a p te r
E R
D is trib u tio n s
LO 5 .5 C o m p u te s u m m a r y m e a s u r e s to e v a lu a te portfo lio re tu rn s.
In this chapter we extend our discussion about probability by introducing the concept o f a random
variable. A random variable summarizes the results o f an experim ent in terms o f numerical values.
It can be classified as discrete or continuous depending on the range o f values tha t it assumes. A
discrete random variable assumes a countable num ber o f distinct values, whereas a continuous ran
dom variable is characterized by infinitely uncountable values. In this chapter, we focus on discrete
random variables. Examples include the num ber o f credit cards carried by consumers, the num ber
o f foreclosures in a sample o f 100 households, and the num ber o f cars lined up at a toll booth. Once
we define the range o f possible values that a random variable assumes, we construct a probability
distribution to com pute the probabilities associated w ith these different values. We also calculate
summary measures for a random variable, including its mean, variance, and standard deviation. Fi
nally, we discuss three w idely used discrete probability distributions: the binomial, the Poisson, and
the hypergeometric distributions.
I N T R O D U C T O R Y C A S E
We generally use the letter X to denote a random variable. A discrete random variable
a s s u m e s a c o u n ta b l e n u m b e r o f d is t in c t v a lu e s s u c h a s x 1,x 2, x 3 a n d s o o n . It m a y a s s u m e e it h e r
a finite number of values, or an infinite sequence of values. A continuous random variable,
on the other hand, is characterized by (infinitely) uncountable values. In other words, a con
tinuous random variable can take on any value within an interval or collection o f intervals.
A discrete random variable assumes a countable num ber o f distinct values, whereas
a continuous random variable is characterized by (infinitely) uncountable values
within any interval.
Recall from Chapter 4, the sam ple space is a set o f all outcom es o f a random experiment.
W henever some numerical values are assigned to these outcom es, a random variable X
is defined. Consider the following experim ents, and some exam ples o f discrete random
variables (with their possible values) that are associated with the experiments:
Experim ent 1. Rolling a six-sided die; sam ple space = {1, 2, 3, 4, 5, 6}.
Let X = the num ber rolled; possible values: {1, 2, 3, 4, 5, 6}
Let X = the odd num ber rolled; possible values = {1, 3, 5}
Experim ent 2. Two shirts are selected from the production line and each can be
defective (D) or non-defective (N); sam ple space = {(D,D), (D,N),
(N,D), (N,N)}.
L e t X = the num ber o f defective shirts; possible values = {0, 1 , 2}
Let X = the proportion o f defective shirts; possible values = {0,1 /2 , 1}
Experim ent 3. Reviewing a single m ortgage application and deciding w hether the
client gets approved (A) or denied (D); sam ple space = {A, D}.
L et X = 1 for A and 0 for D; possible values = { 1 , 0}
L et X = 1 for A and − 1 for D; possible values = { 1 , − 1}
Experim ent 4. Reviewing multiple mortgage applications and, for each client, deciding
whether the client gets approved (A) or denied (D); sample space = the
set o f all possible infinite sequences whose elements are A or D.
Let X = the num ber o f approvals; possible values = {0, 1, 2, 3 . . . .}
Let X = the squared num ber o f approvals; possible values = {0,1, 4, 9 , . . .}
The random variables defined for Experim ents 1, 2 and 3 have finite num bers o f values,
while the two random variables defined for Experim ent 4 have infinite but countable
num bers o f values.
Sometimes, we can define a random variable directly by identifying its values with some
numerical outcomes. For example, we may be interested in the number o f students who
get financial aid out o f the 100 students who applied. Then the set of possible values of the
random variable, equivalent to the sample space, is {0, 1, . . . , 100}. In a similar way, we
can define a discrete random variable with an infinite number of values that it may take. For
example, consider the number of cars that cross the Brooklyn Bridge between 9:00 am and
10:00 am on a Monday morning. Here the discrete random variable takes an infinite but
countable number of values from {0, 1, 2 , . . .}. It is possible that no cars cross the bridge in
this time period; perhaps, due to an accident, the bridge is temporarily closed. Note that we
cannot specify an upper bound on the observed number of cars.
Although, we explore discrete random variables in this chapter, random variables can
also be continuous. For example, the time taken by a student to com plete a 60-minute
exam may assum e any value betw een 0 and 60 minutes. Thus, the set o f such values is
uncountable; that is, it is im possible to put all real num bers from the interval [0, 60] in a
sequence. Here, the random variable is continuous because the outcom es are (infinitely)
uncountable. Some students may think that tim e in the above exam ple is countable in
seconds; however, this is not the case once we consider fractions o f a second. We will
discuss the details o f continuous random variables in the next chapter.
For convenience, in this chapter, we will use terms like “probability distribution” and
“distribution” for the probability mass function. We will do the same in the next chapter
for the probability density function. In both chapters, we will use “cumulative probability
distribution” for the cumulative distribution function.
You can view a discrete probability distribution in several ways, including tabular,
algebraic, and graphical forms. Exam ple 5 . 1 shows one o f two tabular forms. In general,
we can construct a table in two different ways. The first approach directly specifies the
probability that the random variable assum es a specific value.
EXAMPLE 5.1
Refer back to Experim ent 1 o f rolling a six-sided die, with the random variable
defined as the num ber rolled. Present the probability distribution in a tabular form.
SO LU TIO N : A probability distribution for rolling a six-sided die is shown in Table 5.1.
From Table 5.1, we can deduce, for instance, that P(X = 5) equals 1/6. For that m at
ter, the probability that X assumes any o f the six possible values is 1/6 .
T he probability distribution defined in Exam ple 5.1 illustrates tw o com ponents o f all
discrete probability distributions.
T W O KEY P R O P E R T I E S OF D I S C R E T E P R O B A B I L I T Y D I S T R I B U T I O N S
The second tabular view o f a probability distribution is based on the cum ulative prob
ability distribution.
The cumulative probability representation is convenient w hen w e are interested in
finding the probability over a range o f values rather than a specific value. For the ran
dom variable defined in E xam ple 5.1, the cumulative probability distribution is shown in
Table 5.2.
If w e are interested in finding the probability o f rolling a four or less, P(X ≤ 4), we see
from the cumulative probability distribution that this probability is 4 /6 . With the earlier
probability representation, w e would add up the probabilities to com pute P(X ≤ 4) as
At the same time, w e can use the cum ulative probability distribution to find the probabil
ity that the random variable assum es a specific value. For example, P(X = 3) can be found
as P (X ≤ 3) – P(X ≤ 2) = 3 / 6 – 2 / 6 = 1/6.
In many instances w e can express a probability distribution by applying an algebraic
formula. A form ula representation o f the probability distribution o f the random variable
defined in Exam ple 5.1 is:
Thus, from the form ula w e can ascertain that P(X = 5) = 1 /6 and P(X = 7) = 0.
In order to graphically depict a probability distribution, w e place all values x o f X on
the horizontal axis and the associated probabilities P (X = x ) on the vertical axis. We then
draw a line segm ent that em erges from each x and ends w here its height equals P (X = x ).
Figure 5.1 graphically illustrates the probability distribution o f the random variable de
fined in E xam ple 5.1.
EXAMPLE 5.2
Bankrate.com provided the probability distribution in Table 5.3, which reflects the
num ber o f credit cards that its readers carry:
SO LU TIO N:
a. We first note that the random variable X denotes the num ber o f credit
cards that a bankrate.com read er carries. This variable assum es the values
0 through 4. The probability distribution is valid because it satisfies the
follow ing tw o conditions: (1) all percentages fall betw een 0 and 100, or
equivalently, the probabilities fall betw een 0 and 1; and (2) the percentage
sum totals 100 (2.5% + 9.8% + ∙ ∙ ∙ + 54.6% = 100%), or equivalently,
the probability sum totals 1 (0.025 + 0.098 + ∙ ∙ ∙ + 0.546 = 1).
b. In order to find the probability that no bankrate.com readers carry a credit card,
we first write the question using the appropriate probability statement notation.
We find P (X = 0) = 0.025.
c. We express the appropriate probability statement and then sum the
respective probabilities: P(X < 2) = P (X = 0) + P(X = 1) = 0.025 +
0.098 = 0.123.
d. We again write the probability statement and then sum the respective
probabilities: P(X ≥ 2) = P(X = 2) + P (X = 3) + P(P = 4*) = 0.166 +
0.165 + 0.546 = 0.877.
N ote that since the sum o f the probabilities over all values of X equals 1,
we can also find the above probability as P(X > 2) = 1 – P(X < 2) =
1 – 0.123 = 0.877.
e. The graph in Figure 5.2 shows that the distribution is not sym metric, rather it
is skewed to the left. There are small chances o f low values, namely carrying
no m ore than one credit card. The m ost likely value by far is carrying four or
m ore credit cards, w ith a likelihood o f 54.6%.
EXERCISES 5. 1
0.35
x)I
0.10 0.10 control, some argue that the birth rate, especially in rural
India, is still too high to be sustainable. A demographer
a. Complete the probability distribution.
computes the follow ing probability distribution o f the
b. Graphically depict the probability distribution and
household size in India.
com m ent on the symmetry o f the distribution.
c. What is the probability that the random variable X is Household Size Probability
negative? 1 0.05
d. What is the probability tha t the random variable X is 2 0.09
greater than – 20? 3 0.12
e. W hat is the probability tha t the random variable X is less 4 0.24
than 20? 0.25
5
3. Consider the follow ing cum ulative probability distribution. 6 0.12
x 1 2 3 4 5 7 0.07
0
p(X ≤ x) 0.15 0.35 0.52 0.78 0.84 1 8 0.06
a. What is the probability that there are less than 5 c. W h a t is th e p r o b a b ility t h a t he m a kes a t least o n e o f th e
members in a typical household in India? shots?
b. W h a t is th e p r o b a b ility t h a t th e re are 5 o r m o re m e m b e rs d. In a tig h t game, should the opposing team's coach have
in a ty p ic a l h o u s e h o ld in Ind ia? his players foul this player? Explain why or why not.
c. What is the probability that the number of members (A regular basketball shot is worth tw o points, but each
in a typical household in India is strictly between free throw is w orth only one point.)
3 and 6? 8. In e a rly 20 1 0 , le a d in g U.S. s to c k m a rk e ts t u m b le d m o re th a n
d. Graphically depict this probability distribution and 2.5% as U.S. consumer confidence fell to its lowest level
com m ent on its symmetry. since August 2009 (BBC News, July 16, 2010). Given fresh
6. A financial analyst creates the follow ing probability economic data, an economist believes there is a 35% chance
distribution for the performance of an equity income that consumer confidence w ill fall below 62 and only a 25%
mutual fund. chance tha t it w ill rise above 65.
a. According to the economist, what is the probability that
Performance Probability
consumer confidence w ill be between 62 and 65?
Very poor 0.14
b. According to the economist, w hat is the probability that
Poor 0.43
consumer confidence w ill not fall below 62?
Neutral 0.22
9. P rofe ssor S anch ez has b e e n te a c h in g P rin c ip le s o f E cono m ics
Good 0.16 for over 25 years. He uses the follow ing scale for grading.
Very good 0.05
Grade Probability
E X P E C T E D V A L U E OF A D I S C R E T E R A N D O M V A R I A B L E
F o r a d is c re te r a n d o m v a ria b le X w ith v a lu e s x 1, x 2, x 3, . . . th a t o c c u r w ith p r o b a b ilitie s
P (X = x i), th e e x p e c t e d v a lu e o f X is c a lc u la te d as
V A R I A N C E A N D S T A N D A R D D E V I A T I O N OF A D I S C R E T E
R A N D O M VARIABLE
For a discrete random variable X with values x 1, x 2, x3, . . . that occur with probabili
ties P (X = x,) , the variance o f X is calculated as
EXAMPLE 5.3
Brad Williams is the owner o f a large car dealership in Chicago. Brad decides to
construct an incentive compensation program that equitably and consistently com
pensates employees on the basis o f their performance. He offers an annual bonus
o f $10,000 for superior performance, $6,000 for good performance, $3,000 for fair
performance and $0 for poor performance. Based on prior records, he expects an
employee to perform at superior, good, fair, and poor performance levels with prob
abilities 0.15, 0.25, 0.40, and 0.20, respectively. Table 5.4 lists the bonus amount,
performance type, and the corresponding probabilities.
TABLE 5.4 Data fo r Example 5.3
Bonus (in $1,000s) Performance Type Probability
$10 Superior 0.15
$6 Good 0.25
$3 Fair 0.40
$0 Poor 0.20
a . Calculate the expected value o f the annual bonus amount.
b. Calculate the variance and standard deviation o f the annual bonus amount.
c . W hat is the total annual am ount that Brad can expect to pay in bonuses if he
has 25 employees?
SO LU TIO N:
a. Let the random variable X denote the bonus am ount (in $ 1,000s) for an em
ployee. The first and second columns o f Table 5.5 show the probability distri
bution o f X. The calculations o f the m ean are provided in the third column. We
weigh each outcom e by its respective probability, x iP(X = x i), and then sum
these weighted values. Thus, as shown at the bottom o f the third column,
E(X ) = μ = ∑ x i P(X = x ,i) = 4.2, or 4,200. Note that the expected value is
not one o f the possible values o f X ; that is, none o f the em ployees will earn a
bonus o f 4,200. This outcom e reinforces the interpretation o f expected value
as a long-run average.
SO LU TIO N : Let the random variable X represent the worth o f the painting. Given
the above information, we define the probability distribution as shown in Table 5.6.
Since the expected value o f the painting is m ore than $1,000, it m ay appear that
the right choice is to pick the painting over $1,000 in cash. This choice, however,
is based entirely on the expected value o f the painting, ignoring the risk co m
pletely. W hile the expected value o f $1,050 is m ore than $1,000, the painting
entails som e risk. For instance, there is a 30% chance that it m ay be w orth only
$500. The decision to pick the painting m akes no allow ance for risk. In fact, w ith
the above logic, you w ould choose the painting over cash even if the risk w ere
significantly higher.
In general, a risk averse consum er dem ands a positive expected gain as compensation
for taking risk. This com pensation increases w ith the level o f risk taken and the degree of
risk aversion. A risk neutral consum er, on the other hand, com pletely ignores risk and
m akes his/her decisions solely on the basis o f expected values.
In general, consum ers are risk averse and expect a reward for taking risk. A risk
averse consum er may decline a risky prospect even if it offers a positive expected
gain. A risk neutral consum er completely ignores risk and always accepts a pros
pect that offers a positive expected gain. Finally, a risk loving consum er may
accept a risky prospect even if the expected gain is negative.
In E xam ple 5.4, a risk neutral consum er will take the painting because its expected
value exceeds the risk-free cash value o f $1,000. This consum er is not concerned with
risk, as m easured by the standard deviation. A risk lover will be thrilled to take the
painting. For a risk averse consum er, however, the decision is not clear cut. It depends
on the risk involved in picking the painting and how m uch he/she wants to be co m p en
sated for this risk. O ne way w e resolve this issue is to define the utility function o f the
consum er, which in essence conveys the degree o f risk aversion. A risk averse consum er
will pick the risky prospect if the expected utility (not the expected m oney) o f the risky
prospect exceeds the utility o f a risk-free alternative. Further details are beyond the
scope o f this book.
EXERCISES 5.2
Concepts break down w ith a scrap value of $100. Should you buy the
insurance (assume risk neutrality)?
10. Calculate th e mean, variance, and standard d e via tion o f th e
fo llo w in g discrete p ro b a b ility d is trib u tio n . 16. Four years ago, Victor Consuelo purchased a very reliable
3 0.25 10% 17. Market observers are quite uncertain w hether the stock
market has bottom ed out from the economic m eltdown
a. Calculate the expected rate of return. tha t began in 2008. In an interview on March 8,2009,
b. Calculate the variance and standard deviation o f this CNBC interviewed tw o prom inent economists who offered
probability distribution. differing views on whether the U.S. economy was getting
13. Organizers o f an outdoor summer concert in Toronto are stronger or weaker. An investor not wanting to miss out
concerned about the weather conditions on the day o f the on possible investment opportunities considers investing
concert. They w ill m ake a p ro fit o f $25,000 on a clear day and $10,000 in th e stock m arket. He believes th a t th e p ro b a b ility
$10,000 on a cloudy day. They will make a loss of $5,000 if it is 0.30 that the market w ill improve, 0.40 that it w ill stay
rains. The weather channel has predicted a 60% chance of the same, and 0.30 that it will deteriorate. Further, if the
rain on the day o f the concert. Calculate the expected profit economy improves, he expects his investment to grow to
from the concert if the likelihood is 10% tha t it will be sunny $15,000, but it can also go down to $8,000 if the economy
and 30% tha t it w ill be cloudy. deteriorates. If the economy stays the same, his investment
will stay at $10,000.
14. Mark Underwood is a professor o f Economics at Indiana
University. He has been teaching Principles of Economics for a. What is the expected value of his investment?
over 25 years. Professor U n d e rw o o d uses th e fo llo w in g scale b. What should the investor do if he is risk neutral?
fo r grading. c. Is the decision clear cut if he is risk averse? Explain.
year p e rio d th e re is an 8% chance th a t y o u r la p to p c o m p u te r b. Find the expected value and the standard deviation of
w ill require w o rk w o rth $400, a 3% chance th a t it w ill require returns for Fund 2.
work worth $800, and a 2% chance tha t it will completely c. Which fund will you pick if you are risk averse? Explain.
19. Investment advisors recommend risk reduction through State of the Returns Returns
international diversification. International investing allows U.S. Economy in Europe in Asia
you to take advantage o f the potential for growth in foreign
Good 10% 18%
economies, particularly in emerging markets. Janice Wong
Fair 6% 10%
is considering investment in either Europe or Asia. She has
Poor − 6% − 12%
s tu d ie d these m arkets and believes th a t b o th m arkets w ill be
influenced by the U.S. economy, which has a 20% chance for a. Find the expected value and the standard deviation of
being good, a 50% chance for being fair, and a 30% chance returns in Europe and Asia.
for being poor. Probability distributions of the returns for b. W hat w ill Janice pick as an in vestm ent if she is risk neutral?
these markets are given in the accompanying table.
c. Discuss Janice's decision if she is risk averse.
Using algebra, it can be shown that the variance o f the sum o f tw o random variables,
Var(X + Y), yields
For given constants a and b, the above results are extended as:
Note that the portfolio weights add up to one, that is, wA + wB = 0.20 + 0.80 = 1. We
then define the portfolio return Rp as a linear combination o f the individual returns,
P O R T F O L I O E XP EC TE D RETURN
Given a portfolio with two assets, A sset A and Asset B, the expected return o f the
portfolio E (RP) is com puted as
where wA and w B are the portfolio weights (wA + w B = 1) and E(RA) and E (RB) are
the expected returns on assets A and B, respectively.
EXAMPLE 5.5
Consider an investment portfolio o f $40,000 in Stock A and $60,000 in Stock B.
Calculate the expected return of this portfolio based on the information in Table 5.7.
SOLUTION: First w e com pute the portfolio weights. Since $40,000 is invested in
Stock A and $60,000 in Stock B, we compute
Thus, using the form ula for portfolio expected return, we solve:
N ote that the portfolio expected return o f 8.36% is lower than the expected return
o f investing entirely in Stock A with an expected return o f 9.5%, yet higher than the
expected return o f investing entirely in Stock B with an expected return o f 7.6%.
The risk o f the portfolio depends not only on the individual risks o f the assets but also on
the interplay betw een the asset returns. For example, if one asset does poorly, the second
asset may serve as an offsetting factor to stabilize the risk o f the overall portfolio. This
result will work as long as the return o f the second asset is not perfectly correlated with
the return of the first asset. Sim ilar to the covariance Cov(x, y ) = σ xy introduced in C hap
ter 3, the covariance Cov(R A, RB) = σA B helps determine whether the linear relationship
between the asset returns is positive, negative, or zero. Recall that an easier m easure to in
terpret is the correlation coefficient ρ which describes both the direction and the strength
o f the linear relationship betw een two random variables. T he value o f the correlation
coefficient falls between − 1 and 1. The closer the value is to 1, the stronger is the positive
relationship between the variables. Similarly, the closer the value is to − 1, the stronger
is the negative relationship between the variables. Let denote the correlation
coefficient between the returns RA and RB.
With inform ation on either the covariance or the correlation coefficient or the two
returns, w e can now determ ine the portfolio variance o f return.
PORTFOLIO VARIANCE
The portfolio variance, Var(Rp) = Var(w ARA + w BR B), is calculated as
or, equivalently,
where σ2A and σ2B are the variances o f the returns for A sset A and A sset B, respec
tively, σAB is the covariance between the returns for Asset A and Asset B, and ρ AB is
the correlation coefficient between the returns for Asset A and A sset B.
The standard deviation o f return SD(RP) is then calculated as the positive square root
o f the portfolio variance.
EXAMPLE 5 .6
U sing the information in Exam ple 5.5, answ er the following questions.
a. Calculate and interpret the correlation coefficient betw een the returns on
Stocks A and B.
b. Calculate the portfolio variance using both formulas.
c. Calculate the portfolio standard deviation.
d. C om m ent on the findings.
SOLUTION:
a. We calculate the correlation coefficient as
This value implies that the returns have a positive linear relationship, though
the m agnitude o f the relationship is w eak (ρ AB is well below 1).
b. Using the first form ula for portfolio variance, we calculate
EXERCISES 5.3
20. What are the portfolio weights for a portfolio that has 100 m arket fu n d yields a risk-free return o f 4%. The in p u ts fo r th e
shares o f Stock X th a t sell fo r $20 per share and 200 shares o f risky fu n d s are given below.
Stock Y th a t sell fo r $12 per share? Fund Expected Return Standard Deviation
21. You own a portfolio that has $4,400 invested in stocks and Stock fund 14% 26%
$5,600 invested in bonds. W h a t is th e expected return o f th e Bond fund 8% 14%
p o rtfo lio if stocks and bonds are expected to yield a return o f
9% and 5%, respectively? The co rrelation c o e fficie nt b e tw e en th e stock and b ond
fu n d s is 0.20.
22. A portfolio has $200,000 invested in Asset X and $300,000 in
a. W h a t is th e expected return and th e variance o f a
Asset Y. Consider th e sum m ary measures in th e fo llo w in g table.
portfolio that invests 60% in the stock fund and 40% in
Measures Asset X Asset Y
the bond fund?
Expected Return (%) 8 12
b. What is the expected return and the variance of a
Standard deviation (%) 12 20 portfolio tha t invests 60% in the stock fund and 40% in
Correlation 0.40 th e m on e y m arke t fund? [Hint: Note th a t the correlation
between any asset and the risk-free T-bills is zero.]
a. Calculate th e p o rtfo lio w e ig h ts fo r assets X and Y.
c. Com pare th e p o rtfo lio s in parts a and b w ith a p o rtfo lio
b. Calculate th e expected return o f th e p o rtfo lio .
tha t is invested entirely in the bond fund.
c. Calculate th e standard d e v ia tio n o f th e p o rtfo lio .
25. You have $400,000 invested in a well-diversified portfolio.
23. An analyst has pre d icte d th e fo llo w in g returns fo r Stocks A
You inherit a house tha t is presently worth $200,000.
and B in th re e possible states o f th e econom y.
Consider the summary measures in the following table:
State Probability A B
Investment Expected Return Standard deviation
Boom 0.3 0.15 0.25
Old portfolio 6% 16%
Normal 0.5 0.10 0.20
House 8% 20%
Recession ? 0.02 0.01
The correlation coefficient between your portfolio and the
a. W h a t is th e p ro b a b ility o f a recession? house is 0.38.
b. Calculate th e expected return o f Stocks A and B. a. W hat is th e expected return and standard deviation o f your
c. Calculate th e expected return o f a p o rtfo lio th a t is portfolio comprising your old portfolio and the house?
invested 55% in A and 45% in B. b. Suppose you decide to sell the house and use the
24. A pension fu n d m anager is considering th re e m u tu a l funds proceeds o f $200,000 to buy risk-free T-bills th a t prom ise
fo r in ve stm e n t. The firs t one is a stock fu n d , th e second is a a 3% rate of return. Calculate the expected return and
bond fu n d and th e th ird is a m o n e y m arke t fu n d . The m on e y standard deviation of the resulting portfolio.
A binom ial random variable X is defined as the num ber of successes achieved in
the n trials o f a Bernoulli process. A binom ial probability distribution shows the
probabilities associated with the possible values o f the binomial random variable.
We w ill eventually arrive at a general fo rm u la that helps us derive a binom ial probability
d istribution. First, however, w e w ill use a specific exam ple and co n stru ct a p rob ab il
ity tree in order to illustrate the possible outcom es and th eir associated probabilities.
EXAMPLE 5.7
From past experience, a m anager o f an upscale shoe store knows that 85% of her
custom ers will use a credit card w hen m aking purchases. Suppose three customers
are in line to make a purchase.
a. D oes this example satisfy the conditions of a Bernoulli process?
b. C onstruct a probability tree that delineates all possible values and their associ
ated probabilities.
c. U sing the probability tree, derive the binomial probability distribution.
SO LU TIO N :
a. This exam ple satisfies the conditions o f a Bernoulli process because a cus
tom er either uses a credit card (labeled success), with an 85% likelihood, or
does not use a credit card (labeled failure), w ith a 15% likelihood. M oreover,
given a large num ber o f custom ers, these probabilities o f success and failure
do not change from custom er to customer.
b. In Figure 5.3, w e let S denote the outcom e that a custom er uses a credit card
and F denote the outcom e that a custom er does not use a credit card. Start
ing from the unlabeled node on the left, each branch reflects the probability
o f that branch outcom e’s occurring. For instance, there is an 85% chance
that custom er 1 uses a credit card. The branches emanating from custom er
1 denote conditional probabilities o f custom er 2 using a credit card, given
w hether custom er 1 used a credit card. However, since we assum e that the tri
als o f a Bernoulli process are independent, the conditional probability o f the
branch outcom e’s occurring is the same as its unconditional probability. In
other words, custom er 2 has an 85% chance o f using a credit card and a 15%
chance o f not using one. The same holds for the probabilities for custom er 3.
The fourth colum n shows that there are eight possible outcom es at the conclu
sion o f this experiment. We are able to obtain relevant probabilities by using
the multiplication rule for independent events. For instance, following the top
branches throughout the probability tree, we calculate the probability that all
three customers use a credit card as (0.85)(0.85)(0.85) = 0.614. The probabili
ties for the rem aining outcom es are found in a similar manner.
c. Since w e are not interested in identifying the particular custom er who uses a
credit card, but rather the num ber o f customers who use a credit card, we can
com bine events with the same num ber o f successes, using the addition rule for
mutually exclusive events. For instance, in order to find the probability that one
custom er uses a credit card, w e add the probabilities that correspond to the out
come x = 1 (see shaded areas in table): 0.019 + 0.019 + 0.019 = 0.057. Simi
larly, we calculate the rem aining probabilities corresponding to the other values
o f X and construct the probability distribution shown in Table 5.8. N ote that in
many solved problems, the probabilities do not add up to 1 due to rounding.
Fortunately we do not have to construct a binom ial probability tree each time w e w ant to
find a binom ial probability distribution. We can use the following form ula for calculating
the probability associated with a binom ial random variable.
THE B I N O M I A L P R O B A B IL IT Y D I S T R I B U T I O N
For a binomial random variable X , the probability o f x successes in n Bernoulli
trials is
f o r x = 0, 1, 2 , . . . , n . By definition, 0! = 1.
In other words, each sequence consisting o f 1 success in 3 trials has a 1.9% chance of
occurring.
In order to obtain the overall probability o f getting 1 success in 3 trials, we then m ul
tiply the binomial coefficient by the probability o f obtaining the particular sequence, or
here, 3 × 0.019 = 0.057. This is precisely the probability that we found for P (X = 1)
using the probability tree.
Further, w e could use the form ulas shown in Section 5.2 to calculate the expected value,
the variance, and the standard deviation o f any binom ial random variable. Fortunately, for
the binom ial distribution, these form ulas simplify to E (X ) = np and Var(X ) = npq. The
simplified form ula for expected value is rather intuitive in that if we know the probability
o f success p o f a random experim ent and w e repeat the experim ent n times, then on aver
age, w e expect np successes.
EXPE CTE D VALUE, V AR IA N C E , A N D S T A N D A R D D E V I A T I O N
OF A B I N O M I A L R A N D O M V A R I A B L E
If X is a binomial random variable, then
For instance, for the binomial probability distribution assum ed in Exam ple 5.7, we can
derive the expected value with the earlier general form ula as
However, an easier way is to use E(X ) = np and thus calculate the expected value as
3 × 0.85 = 2.55. Similarly, the variance can be easily calculated as
EXAMPLE 5.8
Approxim ately 20% o f U.S. workers are afraid that they will never be able to retire
(bankrate.com, June 23, 2008). Suppose 10 workers are random ly selected.
a. W hat is the probability that none o f the workers is afraid that they will never
be able to retire?
b. W hat is the probability that at least two o f the workers are afraid that they will
never be able to retire?
c. W hat is the probability that no more than two o f the workers are afraid that
they will never be able to retire?
d. Calculate the expected value, the variance, and the standard deviation o f this
binomial probability distribution.
S O LU TIO N : First, this problem satisfies the conditions o f a Bernoulli process. Here,
a worker is either afraid that he/she will never be able to retire, with probability
p = 0.20, or is not afraid, with probability q = 1 – 0.20 = 0.80. In addition, the
random selection o f 10 workers, n = 10, fulfills the requirem ent that the probability
that a worker fears that he/she will never be able to retire stays the same from
worker to worker.
a. We let x = 10 and find
In other words, there is a 10.74% chance that none o f the workers is afraid that
they will never be able to retire.
b. The phrase “at least two w orkers” leads to the following probability statement:
We can solve this problem by first calculating each of the nine probabilities,
from P(X = 2) to P(X = 10). A simpler method uses one o f the key properties
o f a probability distribution, which states that the sum o f the probabilities over
all values o f X equals 1. Therefore, P(X ≥ 2) can be written as 1 – [P (X = 0) +
P(X = 1)], where we need to calculate only two probabilities, P(X = 0) and
P(X = 1), to solve the problem. We know from part (a) that P(X = 0) = 0.1074.
Similarly, we can calculate P (X = 1 ) :
We have already found P(X = 0) from part a and P(X = 1) from part b. So we
now com pute P (X = 2):
d. We use the simplified formulas for these summary m easures and obtain:
EXAMPLE 5.9
In 2007 approxim ately 4.7% o f the households in the D etroit m etropolitan area
were in some stage o f foreclosure, the highest foreclosure rate in the nation ( The
Associated Press, February 13, 2008). Suppose w e sample 100 mortgage-holding
households in the Detroit area.
a. W hat is the probability that exactly 5 o f these households are in some stage of
foreclosure?
b. W hat is the probability that no m ore than 5 o f these households are in some
stage o f foreclosure?
c. W hat is the probability that m ore than 5 households are in some stage o f
foreclosure?
SO LU TIO N :
a. It is possible to use the binom ial form ula and solve this problem as
but we would quickly find the arithmetic
quite unwieldy. Using the binomial function on Excel (Formulas > Insert
Function > BINO M .DIST), we supply the following four arguments in the
dialog box, as shown in Figure 5.4.
• Number_s is the num ber o f successes in n trials. We enter 5.
• Trials is the num ber o f independent trials. We enter 100.
• Probability _s is the probability o f success on each trial. We enter 0.047.
• Cumulative is a logical value. If we enter the value 1 or TRUE, Excel will
return a cumulative probability, or in this case P (X ≤ 5). Since we want
P(X = 5), which is a probability mass function, we enter 0 or FALSE.
FIGURE 5.4 C om puting binom ial probabilities w ith Excel (Example 5.9a)
Excel returns the form ula result as 0.1783; thus, P(X = 5) = 0.1783.
b. We write the probability that no more than 5 o f these households are in some
stage o f foreclosure as P (X ≤ 5). U sing Excel w e input data as shown in
Figure 5.5.
FIGURE 5.5 C om puting cum ulative binom ial probabilities w ith Excel (Example 5.9b)
In this case, the only value that changes is for Cumulative. Here, we enter 1
for cumulative because we are solving for a cum ulative probability. Excel
returns the probability 0.6697; thus, P(X ≤ 5) = 0.6697.
c. The probability that more than five households are in some stage o f foreclo
sure is written as P {X > 5). U sing the information in part b, w e solve this as
P(X > 5) = 1 − P(X ≤ 5) = 1 − 0.6697 = 0.3303.
EXERCISES 5.4
29. Let the probability of success on a Bernoulli trial be 0.30. c. What is the expected num ber tha t w ill graduate?
In five Bernoulli trials, w hat is the probability that there 34. The percentage o f Americans w ho have confidence in U.S.
w ill be (a) 4 failures, (b) more than the expected num ber of banks dropped to 23% in June 2010, which is far below
failures? the pre-recession level o f 41% reported in June 2007
30. (Use com puter) Let X represent a binom ial random (gallup.com).
variable w ith n = 150 and p = 0.36. Find the follow ing a. What is the probability th a t fewer than half of
probabilities. 10 Americans in 2010 have confidence in U.S. banks?
c. P(X > 60) 35. In recent analyses o f Census figures, one in four American
We first m ust ensure that our random experim ent satisfies the conditions o f a Poisson
process.
A random experim ent satisfies a Poisson process if:
• The num ber o f successes within a specified time or space interval equals any
integer between zero and infinity.
• The numbers of successes counted in nonoverlapping intervals are independent.
• The probability that success occurs in any interval is the same for all intervals of
equal size and is proportional to the size o f the interval.
For a Poisson process, w e define the num ber o f successes achieved in a specified tim e or
space interval as a Poisson random variable. Like the Bernoulli process, many random
experim ents fit the conditions o f a Poisson process, for instance:
As w ith the binomial random variable, we have simplified formulas to calculate the vari
ance and standard deviation o f a Poisson random variable. An interesting fact is that the
m ean o f the Poisson random variable is equal to the variance.
EXPECTED VALUE, VA R IA N C E , A N D S T A N D A R D D E V IA T IO N
OF A P O I S S O N R A N D O M V A R I A B L E
If X is a Poisson random variable, then
EXAMPLE 5 .10
We can now address questions first posed by A nne Jones in the introductory case
o f this chapter. Recall that Anne is concerned about staffing needs at the Starbucks
that she manages. She has specific questions about the probability distribution of
custom er arrivals at her store. A nne believes that the typical Starbucks custom er
averages 18 visits to the store over a 30-day month. She wants answers to the fol
lowing probabilities:
a. How many visits should A nne expect in a 5-day period from a typical
Starbucks custom er?
b. W hat is the probability that a custom er visits the chain five times in a 5-day
period?
c. W hat is the probability that a custom er visits the chain no more than two times
in a 5 -day period?
d. W hat is the probability that a custom er visits the chain at least three times in a
5-day period?
c. For the probability that a custom er visits the chain no more than two times in
a 5-day period, w e express the appropriate probability statement as P(X ≤ 2).
Since this probability is equivalent to P(X = 0) + P(X = 1) + P(X = 2), we
first m ust calculate these individual probabilities and then find the sum:
A n n e Jones, th e m a n a g e r o f a S ta rb u c k s store, is c o n c e rn e d a b o u t
h o w o th e r n e a rb y s to re c lo s in g s m ig h t a ffe c t f o o t tr a ffic a t h e r
store. A s o lid u n d e r s ta n d in g o f t h e lik e lih o o d o f c u s to m e r arriv a ls
is n e cessa ry b e fo re she can m a k e f u r t h e r s ta tis tic a l in fe re n c e . H is
to ric a l d a ta a llo w h e r to assu m e t h a t a ty p ic a l S ta rb u c ks c u s to m e r
ave ra g e s 18 v is its t o a S ta rb u c ks s to re o v e r a 3 0 -d a y m o n th . W ith
th is in fo r m a tio n a n d th e k n o w le d g e th a t she can m o d e l c u s to m e r
arriva ls u s in g th e Poisson d is tr ib u tio n , she d e d u c e s t h a t a ty p ic a l
c u s to m e r ave ra g e s th r e e v is its in a 5 -d a y p e rio d . T h e lik e lih o o d
t h a t a ty p ic a l c u s to m e r fre q u e n ts h e r s to re fiv e tim e s in a 5 -d a y
p e rio d is a p p r o x im a te ly 10%. F u rth e r, th e re is a p p r o x im a te ly a
4 2 % c h a n c e t h a t a ty p ic a l c u s to m e r g o e s t o S ta rb u c ks n o m o re
th a n t w o tim e s , w h ile th e c ha nce s t h a t th is c u s to m e r v is its th e c h a in a t le a st th re e tim e s
is a p p r o x im a te ly 58% . These p re lim in a r y p ro b a b ilitie s w ill p ro v e v ita l as A n n e p la n s h e r
f u tu r e s ta ffin g needs.
E X A M P L E 5 .11
Last year, even as a recession gripped the country, 114 m icrobreweries and brew-
pubs opened in the United States ( The Wall Street Journal, M arch 18, 2009). A s
sum e this num ber represents an average and remains constant over time. Solve the
follow ing probabilities with Excel.
a. W hat is the probability that no m ore than 100 microbrew eries or brewpubs
open in a given year?
b. W hat is the probability that exactly 115 m icrobreweries or brew pubs open in a
given year?
SOLUTION:
a. We w ish to determ ine the pro b ab ility that no m ore than 100 m icro brew
eries or brew pubs open in a given year, that is, P(X ≤ 100). U sing the
Poisson function on Excel (F orm u las > Insert F u n ction > P O ISSO N .
D IST ), w e supply three argum ents in the P O IS S O N dialog box as show n
in Figure 5.6.
• X is the num ber o f successes over some interval. We enter 100.
• M ean is the expected value and it m ust be a positive value. We enter 114.
• Cum ulative is a logical value. If w e enter the value 1 or TRUE, Excel will
return a cumulative probability or in this case P(X ≤ 100); If w e enter the
value 0 or FALSE, Excel will return a probability mass function or in this
case P(X = 100). We enter 1.
Excel returns the form ula result as 0.1012; thus, there is approxim ately a 10%
chance that no more than 100 m icrobreweries or brew pubs will open in any
given year.
FIGURE 5.6 C om puting Poisson probabilities w ith Excel (Example 5.11a)
b. Here we wish to find P(X = 115) or the probability that exactly 115 microbrew
eries or brewpubs open in any given year. We input the value 115 for X ,
114 for M ean, and 0 for Cumulative and Excel returns the data in Figure 5.7.
Thus, there is a 3.7% chance that exactly 115 microbreweries or brewpubs will
open in any given year.
EXERCISES 5.5
Concepts 44. Let the mean success rate o f a Poisson process be 8 successes
per hour.
42. Assume that X is a Poisson random variable w ith μ = 1.5.
Calculate the follow ing probabilities. a. Find the expected num ber o f successes in a half-hour
period.
a. P(X = 1)
b. Find the probability o f at least 2 successes in a given
b. P(X = 2)
half-hour period.
c. P(X ≥ 2)
c. Find the expected number o f successes in a tw o-hour
43. Assume tha t X is a Poisson random variable w ith μ = 4. period.
Calculate the follow ing probabilities.
d. Find the probability o f 10 successes in a given tw o-hour
a. P(X = 4)
period.
b. P( X = 2)
45. (Use computer) Assume that X is a Poisson random variable
c. P(X ≤ 1)
w it h μ = 15. Calculate the follow ing probabilities.
a. P(X ≤ 10) 50. Motorists arrive at a Gulf gas station at the rate o f tw o per
b. P(X = 13) m inute during m orning hours.
c. P(X > 15) a. What is the probability that more than tw o motorists
d. P(1 2 ≤ X ≤ 18) w ill arrive at the Gulf gas station during a one-m inute
interval in the morning?
46. (Use computer) Assume t h a t X is a Poisson random variable
b. What is the probability tha t exactly six motorists will
w ith μ = 20. Calculate the follow ing probabilities.
arrive at the Gulf gas station during a five-m inute
a. P(X < 14)
interval in the morning?
b. P(X ≥ 20)
c. How many motorists can an employee expect in her
c. P(X = 25)
three-hour m orning shift?
d. P(1 8 ≤ X ≤ 23)
51. According to a recent governm ent report, the aging of
the U.S. population is translating into many more visits to
Applications
doctors' offices and hospitals (USA Today, August 7 , 2008). It
47. A textile m anufacturing process finds tha t on is estimated tha t an average person makes four visits a year
average, tw o flaws occur per every 50 yards of material to doctors' offices and hospitals.
produced.
a. W hat is the mean and standard deviation o f an average
a. W hat is the probability o f exactly 2 flaws in a 50-yard person's number o f m onthly visits to doctors' offices and
piece o f material? hospitals?
b. What is the probability of no more than tw o flaws in a b. W hat is the probability tha t an average person does
50-yard piece o f material? not make any m onthly visits to doctors' offices and
c. What is the probability of no flaws in a 25-yard piece of hospitals?
material? c. W hat is the probability tha t an average person makes
48. A tollbo oth operator has observed tha t cars arrive randomly at least one m onthly visit to doctors' offices and
at an average rate o f 360 cars per hour. hospitals?
a. Find the probability tha t 2 cars arrive during a specified 52. (Use computer) On average, 400 people a year are
1-m inute period. struck by lightning in the United States (The Boston Globe,
b. Find the probability tha t at least 2 cars arrive during a July 2 1 , 2008).
specified 1-m inute period. a. What is the probability tha t at most 425 people are
c. Find the probability tha t 40 cars arrive between struck by lightning in a year?
10:00 am and 10:10 am. b. What is the probability tha t at least 375 people are struck
49. Airline travelers should be ready to be more flexible as by lightning in a year?
airlines once again cancel thousands o f flights this summer. 53. (Use computer) In the fiscal year tha t ended September 30,
The Coalition for Airline Passengers Rights, Health, and 2008, there were 24,584 age-discrimination claims filed w ith
Safety averages 400 calls a day to help stranded travelers the Equal Employment O pportunity Commission, an increase
deal w ith airlines (seattlepi.com, July 1 0 , 2008). Suppose the of 29% from the previous year (The Wall Street Journal, March
hotline is staffed for 16 hours a day. 7– 8,2009). Assume there were 260 working days in the fiscal
a. Calculate the average num ber o f calls in a one-hour year by which a worker could file a claim.
interval; 30-m inute interval; 15-m inute interval. a. Calculate the average num ber o f claims filed on a
b. What is the probability o f exactly 6 calls in a 15-minute working day.
interval? b. W hat is the probability th a t exactly 100 claims were filed
c. What is the probability o f no calls in a 15-minute on a working day?
interval? c. W hat is the probability tha t no more than 100 claims
d. What is the probability o f at least tw o calls in a were filed on a w orking day?
15-minute interval?
We use the hypergeom etric distribution in place of the binomial distribution when
we are sampling without replacem ent from a population whose size N is not
significantly larger than the sample size n.
In the above example, we assumed sampling without replacement; in other words, after an
item is drawn, it is not put back in the box for subsequent draws. The binomial distribution
would be appropriate if we sample with replacement since, in that case, for each draw there
will be 20 items o f which 2 are defective, resulting in an unchanging probability o f suc
cess. Further, the dependence o f the trials can be ignored if the population size is very large
relative to the sample size. For instance, if the box consists o f 10,000 items o f which 10%,
or 1,000, are defective, then the probability o f success in the second draw will be either
9 9 9/9,999 or 1,000/9,999, which are both approximately equal to 0.10.
Therefore, the likelihood that exactly one out o f five m angoes is dam aged is
39.47%.
b. Note that P (X ≥ 1) = 1 − P(X = 0) where
Therefore, the probability that the shipm ent will be rejected equals P (X ≥ 1) =
1 − P (X = 0) = 1 − 0.5526 = 0.4474.
c. We use the simplified form ulas to obtain
Excel returns the formula result 0.3947 for P (X = 1), which is the value that we obtained
manually.
EXERCISES 5.6
59. Despite the repeated effort by the governm ent to reform 62. The National Science Foundation is fielding applications
how Wall Street pays its executives, some of the nation's for grants to study climate change. Twenty universities
apply for a grant, and only 4 of them w ill be awarded. If six numbers for the base game of Powerball. There are tw o
Syracuse University and Auburn University are among the 20 independent stages o f the game. Five balls are randomly
applicants, w hat is the probability that these tw o universities drawn from 59 consecutively num bered w h ite balls. Further,
will receive a grant? Assume that the selection is made one ball, called the Powerball, is randomly drawn from
randomly. 39 consecutively numbered red balls. To be a winner, the
num bers selected by th e player m ust m atch the num bers on
63. (Use computer) A com m ittee o f 40 members consists of
24 men and 16 women. A subcommittee consisting o f 10 the randomly drawn w hite balls as well as the Powerball.
random ly selected members w ill be formed. a. What is the probability tha t the player is able to match
a. What are the expected num ber of men and women in the numbers of tw o out of five random ly drawn w hite
b. What is the probability tha t at least half o f the members b. What is the probability tha t the player is able to match
in the subcommittee will be women? the numbers of all five random ly drawn w hite balls?
c. What is the probability tha t the player is able to match
64. (Use computer) Powerball is a jackpot game w ith a grand
the Powerball f o r a randomly drawn red ball?
prize starting at $20 million and often rolling over into the
hundreds of millions. In 2006, the jackpot was $365 million. d. What is the probability o f w inning the jackpot? [Hint:
The w in n e r may choose to receive the ja ckp o t prize paid over Remember that the two stages o f drawing white and red
29 years or as a lu m p sum payment. For $1 the player selects balls are independent.]
Probability
Bonus (in $) Grossman Arroyo
0 0.35 0.20
50,000 0.45 0.25
100,000 0.10 0.35
150,000 0.10 0.20
Jordan Lake, an independent hum an resource specialist, is asked to summ arize the
paym ent plans with respect to each executive’s probability distribution.
Jordan would like to use the above probability distributions to:
1. Com pute expected values to evaluate paym ent plans for Grossm an and Arroyo.
2. Help Grossm an and Arroyo decide whether to choose Option 1 or Option 2 for his/
her com pensation package.
Skyhigh Construction, Inc., has ju st hired two new senior executives, A llen Grossm an and Sample
F elicia Arroyo, to oversee planned expansion o f operations. As senior executives, they par
ticipate in a pick-your-salary plan. Each executive is given two options for compensation: Report—
Option 1: A base pay o f $ 125,000 with a possibility of a large bonus.
Comparison of
Option 2: A base pay o f $ 150,000 w ith a possibility o f a bonus, but the bonus would
be one-half o f the bonus under O ption 1. Salary Plans
Grossm an and Arroyo understand that if the firm does not generate an operating profit
in the fiscal year, then no bonuses are paid. Each executive has constructed a probability
distribution given his/her personal background, underlying risk preferences, and subjec
tive view o f the economy.
Given the probability distributions and with the aid o f expected values, the following
analysis will attempt to choose the best option for each executive. Grossman, a married father
with two young children, believes that Table 5.A best reflects his bonus payment expectations.
TABLE 5.A Calculating Grossman's Expected Salary
Bonus (in $) Probability Weighted Value, xi,P(xi)
0 0.35 0 × 0.35 = 0
50,000 0.45 50,000 × 0.45 = 22,500
100,000 0 . 10 100,000 × 0.10 = 10,000
150,000 0. 10 150,000 × 0.10 = 15,000
Total = $47,500
Expected bonus, E(X ), is calculated as a weighted average o f all possible bonus values
and is shown at the bottom o f the third colum n o f Table 1. G rossm an’s expected bonus is
$47,500. U sing this value for his bonus, his salary options are:
Option 1: $125,000 + $47,500 = $ 172,500
Option 2: $ 150,000 + ( 1 /2 × $47,500) = $ 173,750
A rroyo’s expected bonus amounts to $77,500. Thus, her salary options are:
A discrete uniform distribution is a symm etric distribution w here the random vari
able assumes a finite num ber o f specified values and each value is equally likely.
LO 5.3 Calculate and interpret sum m ary m easures for a discrete random variable.
For a discrete random variable X with values x 1, x 2, x3, . . . that occur with probabilities
P (X = xi), the expected value o f X is calculated as We in
terpret the expected value as the long-run average value o f the random variable over in
finitely many independent repetitions of an experiment. M easures o f dispersion indicate
whether the values o f X are clustered about u or widely scattered from μ . The variance
o f X is calculated as The
standard deviation o f X is
LO 5.4 Differentiate a m o n g risk neutral, risk averse, and risk loving consum ers.
In general, a risk averse consum er expects a reward for taking risk. A risk averse
consum er may decline a risky prospect even if it offers a positive expected gain. A risk
neutral consum er completely ignores risk and always accepts a prospect that offers
a positive expected gain. Finally, a risk loving consum er may accept a risky prospect
even if the expected gain is negative.
LO 5.8 Describe the hypergeometric distribution and com pute relevant probabilities.
The hypergeometric probability distribution is appropriate in applications where
the trials are not independent and the probability of success changes from trial to
trial. We use it in place of the binomial distribution when we are sampling with
out replacement from a population whose size N is not significantly larger than
the sample size n. For a hypergeometric random variable X, the probabil
respectively.
a. What is the probability th a t all five flights are delayed? a. W hat is t h e probability th a t a t lea st five p e o p le will show
a. W hat is the probability th a t exactly 15 o f the mortgages a. W hat is th e p robability th a t sh e finds a t least o n e of th e
are underwater? s tu d e n ts ch eatin g ?
b. W hat is the probability th a t more than 20 of the b. W hat is t h e probability th a t sh e finds a t lea st o n e of th e
mortgages are underwater? s tu d e n ts c h ea tin g if sh e fo cu ses o n six random ly ch o sen
72. (Use computer) According to a survey by consulting firm Watson In te rn e t access. S u p p o se 20 o u t of 80 h o u se h o ld s in a small
or plan to do so in the next year (Kiplinger's Personal Finance, th a t provides high sp e e d In tern e t has recently e n te re d th e
February 2009). Suppose 30 employers are randomly selected. m arket. As p a rt of th e m ark etin g cam p aig n , th e co m p an y
d e cid es to random ly se lec t 10 h o u se h o ld s a n d offer th e m
a. What is the probability th a t exactly 10 o f the employers
free la p to p s alo n g w ith a b ro ch u re th a t d e scrib es th eir
have elim inated or plan to elim inate perks?
services. The aim is to build goodw ill and, w ith a free laptop,
b. W hat is the probability th a t at least 10 employers, but
te m p t n o n u se rs into g e ttin g In tern e t access.
no more than 20 employers, have elim inated or plan to
a. W hat is th e probability th a t 6 la p to p recipients d o n o t
elim inate perks?
have In tern e t access?
c. What is the probability th a t at m ost 8 employers have
b. W hat is th e probability th a t a t least five lap to p recipients
elim inated or plan to elim inate perks?
d o n o t have In tern e t access?
73. Studies have shown th a t bats can consume an average o f 10
c. W hat is th e probability th a t tw o or few er lap to p
mosquitoes per m inute (berkshiremuseum.org).
recip ien ts d o n o t have In tern e t access?
a. Calculate the average num ber o f mosquitoes tha t a bat
d. W hat is th e e x p ec te d n u m b e r o f lap to p recipients w h o
consumes in a 30-second interval.
d o n o t have In tern e t access?
CASE STUDIES
Data for Case Study 5.1 Probability Distribution for Repair Cost
LEARNING OBJECTIVES
After reading LO 6.1 Describe a continuous random variable.
this chapter
you should LO 6.2 Describe a continuous uniform distribution and calculate associated probabilities.
be able to:
LO 6.3 Explain the characteristics of the normal distribution.
LO 6.5 Calculate and interpret probabilities for a random variable that follows the
normal distribution.
LO 6.6 Calculate and interpret probabilities for a random variable that follows the
exponential distribution.
LO 6.7 Calculate and interpret probabilities for a random variable that follows the
lognormal distribution.
In the preceding chapter, we defined a random variable and discussed its numerical outcomes. We
then classified the random variable as a discrete or a continuous random variable, depending on
the range o f numerical values that it can assume. A discrete random variable assumes a countable
num ber o f distinct values, such as the num ber o f credit cards carried by consumers, the num ber
o f foreclosures in a sample o f 100 households, and the num ber o f cars lined up at a toll booth. A
continuous random variable, on the other hand, is characterized by (infinitely) uncountable values,
such as the investm ent return on a mutual fund, the w aiting tim e at a toll booth, and the a m ount o f
soda in a cup. In this chapter w e focus our attention on continuous random variables. Most o f this
chapter is devoted to the discussion o f the normal distribution, which is the most extensively used
continuous probability distribution and is the cornerstone o f statistical inference. Other im p o rtan t
continuous distributions discussed in the chapter are the uniform, the exponential, and the lognor
mal distributions.
I N T R O D U C T O R Y CASE
3. D e te r m in e t h e a m o u n t o f s a lm o n t h a t s h o u ld be b o u g h t d a ily so t h a t it m e e ts d e m a n d o n 90 % o f t h e days.
W e p ro v id e a s yno psis o f th is case a t t h e e n d o f S e c tio n 6.3.
6.1 Continuous Random Variables and
the Uniform Probability D istribution
LO 6.1 As discussed in Chapter 5, a discrete random variable X assumes a countable number
Describe a of distinct values such as x 1, x 2, x3, and so on. A continuous random variable, on the other
continuous hand, is characterized by infinitely uncountable values and can take on any value within
random variable. an interval. Unlike the case of a discrete random variable, we cannot describe the pos
sible values of a continuous random variable X with a list x 1, x2, . . . because the outcome
(x1 + x2) / 2 , not in the list, might also be possible.
For a discrete random variable, we can compute the probability that it assumes a
particular value x, or written as a probability statement, P(X = x ). For instance, for a
binomial random variable, we can calculate the probability of exactly one success in n
trials, that is, P(X = 1). We cannot make this calculation with a continuous random vari
able. The probability that a continuous random variable assumes a particular value x is
zero, that is, P(X = x ) = 0. This occurs because we cannot assign a nonzero probability
to each of infinitely uncountable values and still have the probabilities sum to one. Thus,
for a continuous random variable it is only meaningful to calculate the probability that the
value of the random variable falls within some specified interval. Therefore, for a continu
ous random variable, P(a ≤ X ≤ b ) = P(a < X < b ) = P(a ≤ X < b) = P(a < X ≤ b)
since P(X = a) and P(X = b) are both zero.
For a continuous random variable, the counterpart to the probability mass function is
called the probability density function, denoted b y f(x ). As mentioned in Chapter 5, in
this book we often use the term “probability distribution” to refer to both functions. The
graph of f( x ) approximates the relative frequency polygon for the population. Unlike the
discrete probability distribution, f( x ) does not provide probabilities directly. The prob
ability that the variable assumes a value within an interval, say P(a ≤ X ≤ b), is defined
as the area under f ( x ) between points a and b. Moreover, the entire area under f[x) over
all values of x must equal one; this is equivalent to the fact that, for discrete random vari
ables, the probabilities add up to one.
THE P R O B A B I L I T Y D E N S I T Y F U N C T I O N
The probability density function f (x) of a continuous random variable X describes
the relative likelihood that X assumes a value within a given interval, where
• f{x ) > 0 for all possible values x of X , and
• the area under f( x ) over all values of x equals one.
As in the case of discrete random variables, we can use the cumulative distribution
function, denoted by F(x), to compute probabilities for continuous random variables. For
a value x of the random variable X, F(x) = P(X ≤ x) is simply the area under the prob
ability density function/(jc) up to the value x.
THE C U M U L A T IV E D IS T R IB U T I O N F U N C T I O N
For any value x of the random variable X, the cumulative distribution function F(x)
is computed as
If you are familiar with calculus, then you will recognize that this cumulative probability
is the integral o f f(x) in the range below x. Similarly, P(a ≤ X ≤ b) = F(b) — F(a) is
the integral of f(x) between points a and b. Fortunately, we do not necessarily need the
knowledge of integral calculus to compute probabilities with the continuous random vari
ables discussed in this text.
The Continuous Uniform Distribution LO 6 .2
One of the simplest continuous probability distributions is called the continuous Describe a
uniform distribution. This distribution describes a random variable that has an equally continuous uniform
likely chance of assuming a value within a specified range. For example, suppose you distribution and
are informed that your new refrigerator will be delivered between 2:00 pm and 3:00 pm. calculate associated
Let the random variable X denote the delivery time of your refrigerator. This variable is probabilities.
bounded below by 2:00 pm and above by 3:00 pm for a total range of 60 minutes. It is rea
sonable to infer that the probability of delivery between 2:00 pm and 2:30 pm equals 0.50
(= 3 0 /6 0 ), as does the probability of delivery between 2:30 pm and 3:00 pm. Similarly,
the probability of delivery in any 15-minute interval equals 0.25 (= 15/60), and so on.
Figure 6.1 depicts the probability density function of the continuous uniform ran
dom variable. The values a and b on the horizontal axis represent its lower and upper
limits, respectively. The continuous uniform distribution is symmetric around its
mean /jl computed as In the refrigerator delivery example, the mean is computed
implying that you expect the delivery at 2:30 pm. The standard devia
tion σ of a continuous uniform variable equals
It is important to emphasize that the height of the probability density function f ( x ) does
not directly represent a probability. As in the case of all continuous random variables, it
is the area under f ( x ) that corresponds to probability. For the continuous uniform distri
bution, the probability is essentially the area of a rectangle, which is the base times the
height. Therefore, the probability is easily computed by multiplying the length of a speci
fied interval (base) with (height).
THE C O N T I N U O U S U N I F O R M D I S T R I B U T I O N
A random variable X follows the continuous uniform distribution if its probability
density function is
where a and b represent the lower and upper limits of values, respectively, that the
random variable assumes.
The expected value and the standard deviation of X are computed as
EXAMPLE 6.1
A manager of a local drugstore is projecting next month’s sales for a particular cos
metic line. She knows from historical data that sales follow a continuous uniform
distribution with a lower limit of $2,500 and an upper limit of $5,000.
a. What are the mean and standard deviation of this uniform distribution?
b. What is the probability that sales exceed $4,000?
c. What is the probability that sales are between $3,200 and $3,800?
SOLUTION:
a. With a value for the lower limit of a = $2,500 and a value for the upper limit
of b = $5,000, we calculate the mean and standard deviation of this continuous
uniform distribution as
b. We find P(X > 4,000), which is the area between $4,000 and $5,000 in the
graph of the distribution (see Figure 6.2). The base of the rectangle equals
5,000 — 4,000 = 1,000 and the height equals Thus
P{X > 4,000) = 1,000 X 0.0004 = 0.40.
3. For a continuous random variable X, P[20 ≤ X ≤ 40) = 0.15 c. A local carnival is not able to operate its rides if the average
and P(X > 4 0 ) = 0.16. Calculate the follow ing probabilities. price o f electricity is more than 14 cents per kilowatt-hour.
What is the probability that the carnival w ill need to close?
a. P(X < 40)
b. P(X < 20) 9. The arrival tim e o f an elevator in a 12-story dorm itory is
equally likely at any tim e range during the next 4 minutes.
c. P(X = 40)
a. Calculate the expected arrival time.
4. A random variable X follows the continuous uniform
b. What is the probability that an elevator arrives in less
distribution with a lower bound o f 5 and an upper bound o f 35.
than 11/2 minutes?
a. What is the height o f the density function f(x)l
c. What is the probability tha t the w a it for an elevator is
b. What is the mean and the standard deviation of the
more than 11/2 minutes?
distribution?
10. The Netherlands is one o f the world leaders in the production
c. Calculate P(X > 10).
and sale o f flowers. Suppose the heights o f the tulips in the
5. A random variable X follows the continuous uniform
greenhouse o f Rotterdam's Fantastic Flora follow a continuous
distribution w ith a lower bound o f - 2 and an upper
uniform distribution with a lower bound o f 7 inches and an
bound o f 4.
upper bound of 16 inches. You have come to the greenhouse
a. What is the height o f the density function f(x)? to select a bouquet o f tulips, but only tulips w ith a height
b. What are the mean and the standard deviation o f the greater than 10 inches may be selected. What is the probability
distribution? that a randomly selected tulip is tall enough to pick?
c. Calculate P(X ≤ —1). 11. The scheduled arrival tim e for a daily flig h t from Boston to
6. A random variable X follows the continuous uniform New York is 9:25 am. Historical data show that the arrival tim e
distribution w ith a lower lim it o f 10 and an upper lim it o f 30. follows the continuous uniform distribution w ith an early
a. Calculate the mean and standard deviation o f the arrival tim e o f 9:15 am and a late arrival tim e o f 9:55 am.
distribution. a. Calculate the mean and standard deviation of the
b. What is the probability tha t X is greater than 22? distribution.
c. What is the probability tha t X is between 15 and 23? b. What is the probability that a flig h t arrives late?
THE N O R M A L D IS T R IB U T I O N
A random variable X with mean μ and variance σ 2 follows the normal distribution
if its probability density function is
A graph depicting the normal probability density function is often referred to as the
normal curve or the bell curve. The following example relates the normal curve to the
location and the dispersion of the normally distributed random variable.
EXAMPLE 6.2
Suppose we know that the ages of employees in Industries A, B, and C are normally
distributed. We are given the following information on the relevant parameters:
SOLUTION: Figure 6.4 illustrates the difference in location given that the mean age
of employees of Industry A is greater than that of Industry B. Both distributions
show the same dispersion since the standard deviation is the same. Figure 6.5
compares the dispersion given that the standard deviation of age in Industry A is
less than that of Industry C. Here, the peak of Industry A is higher than the peak of
Industry C, reflecting the fact that an employee’s age is likelier to be closer to the
mean age in Industry A. These graphs also serve to point out that we can capture the
entire distribution of any normally distributed random variable based on its mean
and variance (or standard deviation).
We first focus on reading the z table. In the next section, we will show that any normal dis
tribution is equivalent to a standard normal distribution when the unit of measurement is
changed to measure standard deviations from the mean. Therefore, while most real-world
normally distributed variables are not standard normal, we can always transform them
into standard normal and use the z table to compute the relevant probabilities.
Figure 6.6 represents a standard normal or z distribution. Since the random variable
Z is symmetric around its mean of zero, P(Z < 0) = P(Z > 0) = 0.5. As is the case
with all continuous random variables, we can also write the probabilities as P(Z ≤ 0) =
P(Z ≥ 0) = 0.5.
The first column of the table, denoted as the z column, shows values of z up to the tenth
decimal point, while the first row of the table, denoted as the z row, shows hundredths
values. Thus, for z = 1.52, we match 1.5 on the z column with 0.02 on the z row to find a
corresponding probability of 0.9357. The arrows in Table 6.1 indicate that P(Z≤ 1.52) =
0.9357.
In Figure 6.7, the cumulative probability corresponding to z = 1.52 is highlighted.
Note that P(Z ≤ 1.52) = 0.9357 represents the area under the z curve to the left of
1.52. Therefore, the area to the right of 1.52 can be computed as P(Z > 1.52) =
1 – P { Z ≤ 1.52) = 1 – 0.9357 = 0.0643.
F IG U R E 6 .7 Cumulative pro ba bility w ith respect to z = 1.52
- 1 .9 → → → → → → 0.0250
EXAMPLE 6.3
Find the following probabilities for a standard normal random variable Z.
a. P(0 ≤ Z ≤ 1.96)
b. p (1 .5 2 ≤ Z ≤ 1.96)
c. P( - 1.52 ≤ Z ≤ 1.96)
d. P(Z > 4)
FIGURE 6.10 Finding the probability betw een 1.52 and 1.96
d. P(Z > 4) = 1 — P(Z ≤ 4). However, the z table only goes up to 3.99 with
PiZ < 3.99) = 0.9999. For any z value greater than 3.99, it is acceptable to
treat P(Z < z) = 1.0. Therefore, P(Z > 4) = 1 — P(Z ≤ 4) = 1 - 1 = 0
(approximately).
EXAMPLE 6.4
For a standard normal variable Z, find the z values that satisfy the following.
a. P(Z ≤ z) = 0.6808
b. P(Z ≤ z) = 0.90
c. P(Z ≤ z) = 0.0643
d. P ( Z > z ) = 0.0212
e. P { - z ≤ Z ≤ z) = 0.95
0.4 ← ← ← ← ← ← ← 0.6808
b. When deriving z for P(Z ≤ z) = 0.90, we find that the z table (right-hand side)
does not contain the cumulative probability 0.90. In such cases, we use the
closest cumulative probability to solve the problem. Therefore, z is approxi
mately equal to 1.28, which corresponds to a cumulative probability of 0.8997.
Figure 6.13 shows this result graphically.
c. As shown in Figure 6.14, the z value that solves P{Z ≤ z) = 0.0643 must be
negative because the probability to its left is only 0.0643 (less than 0.50). We
look up the cumulative probability 0.0643 (left-hand side) to get z = —1.52.
FIGURE 6.14 Finding z given P (Z ≤ z) = 0.0643
d. We have to find a z value such that the probability to the right of this value is
0.0212. Since the table states cumulative probabilities, we look up P(Z ≤ z) =
1 – 0.0212 = 0.9788 in the table (right-hand side) to g e t z = 2.03. Figure 6.15
shows the results.
e. Since we know that the total probability under the curve equals one, and we
want to find – z and z such that the area between the two values equals 0.95,
we can conclude that the area in either tail is 0.025, that is, P(Z < —z) =
0.025 and P{Z > z) = 0.025. Figure 6.16 shows these results. We then use the
cumulative probability, P(Z ≤ z) = 0.975, to find z = 1.96.
• 95.44% of the values fall within 2 standard deviations of the mean, that is,
P (π – 2 σ ≤ X ≤ μ + σ ) = 0.9544, and
• 99.72% of the values fall within 3 standard deviations of the mean, that is,
P(μ - 3σ ≤ X ≤ μ + 3σ ) = 0.9972.
EXAMPLE 6.5
An investment strategy has an expected return of 4% and a standard deviation of
6%. Assume that investment returns are normally distributed. Use the empirical rule
to answer the following questions.
a. What is the probability of earning a return greater than 10%?
b. What is the probability of earning a return less than – 8%?
EXERCISES 6.2
Concepts 15. Find the following z values for the standard normal variable Z.
a. P(Z ≤ z) = 0.9744
12. Find the follow ing probabilities based on a standard normal
variable Z. b. P (Z > z ) = 0.8389
b. P( Z ≤ -1 .3 2 ) d. P (0 ≤ Z ≤ z) = 0.3315
c. P( 1 . 3 2 < Z < 2 . 3 7 ) 16. Find the following z values for the standard normal variable Z.
13. Find the following probabilities based on a standard normal b. P(z≤ Z ≤ 0) = 0.1772
b. P(Z ≤ -1 .9 2 )
c. P( 0 ≤ Z ≤ 1.62)
Applications
d. P( - 0 . 9 0 ≤ Z ≤ 2.94)
17. The historical returns on a balanced portfolio have had
14. Find the follow ing probabilities based on a standard normal an average return o f 8% and a standard deviation of 12%.
variable Z. Assume tha t returns on this portfolio follow a normal
a. P( - 0 . 6 7 ≤ Z ≤ -0 .2 3 ) distribution. Use the empirical rule to answer the follow ing
b. P( 0 ≤ Z ≤ 1.96) questions.
c. P( - 1 . 2 8 ≤ Z ≤ 0) a. W hat percentage of returns were greater than 20%?
d. P( Z > 4.2) b. What percentage of returns were below - 1 6%?
18. A ssum e th a t IQ scores follow a norm al d istrib u tio n w ith 20. A professional basketball team averages 80 points per game
a m ea n o f 100 a n d a sta n d a rd d eviation of 16. Use th e w ith a standard deviation o f 10 points. Assume points per
em pirical rule to a n sw e r th e follow ing q u estions. game follow the normal distribution. Use the empirical rule
a. W hat p e rc e n ta g e o f p e o p le score b e tw e e n 84 a n d 116? to answer the follow ing questions.
b. W hat p e rc e n ta g e o f p e o p le score less th a n 68? a. What percentage o f scores are between 60 and
100 points?
19. The average rent in a city is $1,500 per m o n th w ith a standard
deviation of $250. A ssum e rent follows th e norm al distribution. b. What percentage o f scores are more than 100 points?
Use th e empirical rule to answ er th e following questions. If there are 82 games in a regular season, in how many
games w ill the team score more than 100 points?
a. W hat p e rce n ta g e of rents are b e tw ee n $1,250 a n d $1,750?
b. W hat p e rc e n ta g e o f re n ts are less th a n $1,250?
c. W hat p e rc e n ta g e o f re n ts are g re a te r th a n $2,000?
THE N O R M A L T R A N S F O R M A T I O N : C O N V E R T I N G X INTO Z
Any normally distributed random variable X with mean μ and standard deviation σ
can be transformed into the standard normal random variable Z as
This normal transformation implies that any value X of X has a corresponding value
z of Z given by
z value has a simple interpretation. It specifies by how many standard deviations the cor
responding * value falls above (z > 0) or below (z < 0) the mean μ . In particular:
• A positive z indicates by how many standard deviations the corresponding x lies above μ
• A zero z indicates that the corresponding x equals μ
EXAMPLE 6.6
Scores on a management aptitude exam are normally distributed with a mean of 72
and a standard deviation of 8.
a. What is the probability that a randomly selected manager will score above 60?
b. What is the probability that a randomly selected manager will score between
68 and 84?
SOLUTION: Let X represent scores with μ = 72 and σ = 8. We will use the normal
transformation to solve these problems.
a. The probability that a manager scores above 60 is P ( X > 60). Figure 6.21 shows
the probability as the shaded area to the right of 60. We transform x = 60 into
Therefore, P{X > 60) = P(Z > -1 .5 ). Since P(Z > -1 .5 ) =
1 - P(Z ≤ – 1.5), we look up – 1.50 in the z table (left-hand side) to get this
probability as 1 – 0.0668 = 0.9332.
FIGURE 6.21 Finding P(X > 60)
b. Here, we find p (68 ≤ X ≤ 84). The shaded area in Figure 6.22 shows this
probability. We transform x = 68 into and x = 84 into
Therefore, P(68 ≤ X ≤ 84) = P ( - 0 .5 ≤ Z ≤ 1.5). We
compute this probability using the z table as P(Z < 1.5) - PiZ. < - 0 .5 )
= 0.9332 – 0.3085 = 0.6247.
THE INVERSE T R A N S F O R M A T I O N : C O N V E R T I N G Z IN T O X
A standard normal variable Z can be transformed to the normally distributed ran
dom variable X with mean /jl and standard deviation a as X = μ + Zσ .
Therefore any value z of Z has a corresponding value x of X given by x = μ + zσ .
EXAMPLE 6.7
Scores on a management aptitude examination are normally distributed with a mean
of 72 and a standard deviation of 8.
a. What is the lowest score that will place a manager in the top 10% (90th per
centile) of the distribution?
b. What is the highest score that will place a manager in the bottom 25% (25th
percentile) of the distribution?
b. We find x such that P{X < x) = 0.25. Using the z table (left-hand side), we find
the corresponding z value that satisfies P(Z < z) = 0.25 as – 0.67. We then
solve x = 72 – 0.67 X 8 = 66.64. Therefore, a score of 66.64 or lower will
place a manager in the bottom 25% of the distribution, (see Figure 6.24).
EXAMPLE 6.8
We can now answer the questions first posed by Akiko Hamaguchi in the introductory
case of this chapter. Recall that Akiko is concerned about buying the right amount of
salmon for daily consumption at Little Ginza. Akiko has estimated that the daily con
sumption of salmon is normally distributed with a mean of 12 pounds and a standard
deviation of 3.2 pounds. She wants to answer the following questions:
a. What proportion of days was the demand at Little Ginza above her earlier
purchase of 20 pounds?
b. What proportion of days was the demand at Little Ginza below 15 pounds?
c. How much salmon should she buy so that it meets customer demand on 90%
of the days?
SOLUTION: Let X denote consumer demand for salmon at the restaurant. We know
that X is normally distributed with μ = 12 and σ = 3.2.
a.
b.
c. In order to compute the required amount of salmon, we solve for x in
P(X ≤ x) = 0.90. Since P(X ≤ x ) = 0.90 is equivalent to P(Z ≤ z) = 0.90, we
first derive z = 1.28. G ivenx = μ + zσ , we findx = 12 + 1.28(3.2) = 16.10.
Therefore, Akiko should buy 16.10 pounds of salmon daily to ensure that
customer demand is met on 90% of the days.
S Y N O P S I S OF I N T R O D U C T O R Y C A S E
FIGURE 6.25 Using Excel to com pute a normal cumulative probability for a given x
• Standard_dev is the standard deviation of the distribution. We enter 3.2.
• Cumulative is a logical value. When we enter the value 1 or TRUE, Excel returns a
cumulative probability.
Excel returns a cumulative probability, or in this case P(X ≤ 20) = 0.9938. Since we want
P(X > 20), we compute 1 — 0.9938 = 0.0062.
In order to compute the required amount of salmon that is needed to satisfy demand on
90% of the days (Example 6.8c), we select Formulas > Insert Function > NORM.INV
from Excel’s menu. In the dialog box shown in Figure 6.26, we supply three arguments in
the NORM.INV dialog box:
• Probability is a cumulative probability associated with the normal distribution. We
enter 0.90.
• Mean is the mean of the distribution. We enter 12.
• Standard_dev is the standard deviation of the distribution. We enter 3.2.
FIGURE 6.26 Using Excel to com pute x for a given normal cum ulative probability
Excel returns the value 16.10. In other words, 16.10 pounds of salmon are needed to meet
customer demand on 90% of the days.
EXERCISES 6.3
23. Let X be normally distributed w ith mean = 120 and
Concepts μ
c. Find x such that P(2500 ≤ X ≤ x) = 0.1217. 31. Loans that are 60 days or more past due are considered
d. Find x such that P(X ≤ x) = 0.4840. seriously delinquent. The Mortgage Bankers Association
reported tha t the rate o f seriously delinquent loans has an
26. The random variable X is normally distributed. Also, it is
average o f 9.1% (The Wall Street Journal, August 2 6 , 2010).
known t h a t P ( X > 150) = 0.10.
Let the rate o f seriously delinquent loans follow a normal
a. Find the population mean μ if the population
distribution w ith a standard deviation o f 0.80%.
standard deviation σ = 15.
a. What is the probability that the proportion o f seriously
b. Find the population mean μ if the population
delinquent loans is above 8%?
standard deviation σ = 25.
b. What is the probability that the proportion o f seriously
c. Find the population standard deviation σ if the
delinquent loans is between 9.5% and 10.5%?
population mean μ = 136.
32. The tim e required to assemble an electronic com ponent is
d. Find the population standard deviation σ if the
normally distributed w ith a mean and standard deviation of
population mean μ = 128.
16 minutes and 8 minutes, respectively.
Applications a. Find the probability tha t a randomly picked assembly
27. The average high-school teacher annual salary is takes between 10 and 20 minutes.
$43,000 (Payscale.com, August 20, 2010). Let teacher b. It is unusual for the assembly tim e to be above
salary be norm ally distributed w ith a standard deviation 24 minutes or below 6 minutes. What proportion of
of $18,000. assembly times fall in these unusual categories?
a. W hat percent of high school teachers make between 33. Recent research suggests that Americans make an average of
$40,000 and $50,000? 10 phone calls per day (CNN, August 26, 2010). Let the number of
b. What percent of high school teachers make more than calls be normally distributed with a standard deviation o f 3 calls.
$80,000? a. What is the probability tha t an average American makes
28. Americans are increasingly skimping on their sleep (National between 4 and 12 calls per day?
Geographic News, February 2 4 , 2005). A health expert b. What is the probability that an average American makes
believes that American adults sleep an average of 6.2 hours more than 6 calls per day?
on weekdays with a standard deviation of 1.2 hours. To c. W hat is the probability tha t an average American makes
answer the following questions, assume that sleep tim e on more than 16 calls per day?
weekdays is normally distributed.
34. The manager o f a night club in Boston stated that 95% o f
a. What percent of American adults sleep more than the customers are between the ages o f 22 and 28 years. If
8 hours on weekdays? the age o f customers is normally distributed w ith a mean o f
b. What percent of American adults sleep less than 6 hours 25 years, calculate its standard deviation.
on weekdays?
35. An estimated 1.8 m illion students take on student loans to
c. What percent of American adults sleep between 6 to pay ever-rising tuitio n and room and board (New York Times,
8 hours on weekdays? April 17,2009). It is also known tha t the average cumulative
29. The w eight of turkeys is normally distributed with a mean of debt o f recent college graduates is about $22,500. Let
22 pounds and a standard deviation of 5 pounds. the cumulative debt among recent college graduates be
a. Find the probability that a randomly selected turkey normally distributed w ith a standard deviation o f $7,000.
weighs between 20 and 26 pounds. Approximately how many recent college graduates have
weighs below 12 pounds. 36. Scores on a marketing exam are known to be normally distributed
30. According to the Bureau of Labor Statistics, it takes an with mean and standard deviation of 60 and 20, respectively.
average of 22 weeks for someone over 55 to find a new job, a. Find the probability that a randomly selected student
compared with 16 weeks for younger workers (The W all Street scores between 50 and 80.
Journal, September 2 , 2008). Assume that the probability b. Find the probability tha t a randomly selected student
distributions are normal and that the standard deviation is scores between 20 and 40.
2 weeks for both distributions. c. The syllabus suggests that the top 15% o f the students
a. What is the probability that it takes a worker over the w ill get an A in the course. What is the m inim um score
age of 55 more than 19 weeks to find a job? required to get an A?
b. What is the probability that it takes a younger worker d. W hat is the passing score if 10% o f the students w ill fail
more than 19 weeks to find a job? the course?
37. Average talk tim e between charges of a cell phone is condominiums in the next quarter be normally distributed
advertised as 4 hours. Assume tha t talk tim e is normally w ith a standard deviation of $15,000.
distributed w ith a standard deviation of 0.8 hour. a. What is the probability tha t the condom inium w ill sell at
a. Find the probability that talk tim e between charges a price (i) below $200,000?, (ii) above $240,000?
for a randomly selected cell phone is below 3.5 hours. b. The company is also trying to sell an artist's condo.
b. Find the probability that talk tim e between charges Potential buyers will find the unusual features o f this
for a random ly selected cell phone is either more than condo either pleasing or objectionable. The manager
4.5 hours or below 3.5 hours. expects the average sale price of this condo to be the
c. Twenty-five percent of the time, talk tim e between same as others at $210,000, but w ith a higher standard
charges is below the 1st quartile value. What is this deviation o f $20,000. W hat is the probability tha t this
value? condo will sell at a price (i) below $200,000?, (ii) above
$240,000?
38. A young investment manager tells his client that the
probability o f making a positive return w ith his suggested 40. You are considering the risk-return profile o f tw o m utual
portfolio is 90%. What is the risk (standard deviation) that funds for investm ent. The relatively risky fun d promises
this investment manager has assumed in his calculation if it an expected return o f 8% w ith a standard deviation of
is known tha t returns are normally distributed w ith a mean 14%. The relatively less risky fund promises an expected
of 5.6%? return and standard deviation o f 4% and 5%, respectively.
Assume th a t the returns are approxim ately norm ally
39. A construction company in Naples, Florida, is struggling to
distributed.
sell condominiums. In order to attract buyers, the company
has made numerous price reductions and better financing a. Which mutual fund w ill you pick if your objective is to
offers. Although condom inium s were once listed for m inim ize the probability o f earning a negative return?
$300,000, the company believes tha t it will be able to get b. Which mutual fund w ill you pick if your objective is to
an average sale price of $210,000. Let the price of these maximize the probability of earning a return above 8%?
where A is a rate parameter and e ≈ 2.718 is the base of the natural logarithm.
The mean and the standard deviation of X are equal: E(X) = SD(X) = 1/λ The
cumulative distribution function of X is
The exponential distribution is based entirely on one parameter, λ > 0 (λ is the Greek
letter lambda), which is often called the rate parameter. The graphs in Figure 6.27 show
the shapes of the exponential distribution based on various values of λ .
FIGURE 6.27 Exponential pro ba bility density fun ction for various
values o f λ
EXAMPLE 6.9
Let the time between e-mail messages during work hours be exponentially distrib
uted with a mean of 25 minutes.
a. Calculate the rate parameter λ .
b. What is the probability that you do not get an e-mail for more than one hour?
c. What is the probability that you get an e-mail within 10 minutes?
SOLUTION:
a. Since the mean E(X) equals
b. The probability that you do not get an e-mail for more than an hour is
P(X > 60). Since P(X ≤ x) = 1 – e -λx, we have P(X > x) = 1 – P(X ≤ x) =
e- λx. Therefore, P(X > 60) = e -0 04(60) = e- 2A0 = 0.0907. The probability
of not getting an e-mail for more than one hour is 0.0907.
c. Here, P(X ≤ 10) = 1– e -0.04(10) = 1 - e -0.04 = 1 - 0.6703 = 0.3297. The
probability of getting an e-mail within 10 minutes is 0.3297.
FIGURE 6.28 Using Excel to com pute an exponential cum ulative pro b a b ility fo r a given x
Excel’s formula result indicates that P(X ≤ 60) = 0.9093. Thus, P (X > 60) = 1 – 0.9093 =
0.0907.
THE L O G N O R M A L D IS T R IB U T I O N
Let X be a normally distributed random variable with mean μ and standard de
viation σ . The random variable Y = ex follows the lognormal distribution with a
probability density function as
The graphs in Figure 6.29 show the shapes of the lognormal density function based on
various values of σ . The lognormal distribution is clearly positively skewed for σ > 1. For
σ < 1, the lognormal distribution somewhat resembles the normal distribution.
F IG U R E 6 , 2 9 Lognormal probability density function for various
values o f a along with n = 0
The mean and the variance of the lognormal random variable Y are related to the mean
and the standard deviation of the corresponding normal random variable X.
E X P E C T E D V A L U E S A N D S T A N D A R D D E V I A T I O N S OF
THE L O G N O R M A L A N D N O R M A L D IS T R IB U T IO N S
Let X be a normal random variable with mean μ and standard deviation σ and
let Y = ex be the corresponding lognormal variable. The mean μ Yand standard
deviation σ Yof Y are derived as
Equivalently, the mean and standard deviation of the normal variable X = ln(Y) are
derived as
EXAMPLE 6.10
Compute the mean and standard deviation of a lognormal random variable if the
mean and the standard deviation of the underlying normal random variable are as
follows:
a. μ = 0, σ = 1
b. μ = 2, σ = 1
c. μ = 2, σ = 1.5
a.
b.
c.
The popularity of the lognormal distribution is also due to the fact that the probabilities of
a lognormal random variable are easily evaluated by reference to the normal distribution.
This is illustrated in the following example.
EXAMPLE 6.11
Let Y = ex where X is normally distributed with mean μ = 5 and standard deviation
σ = 1.2.
SOLUTION: We solve these problems by first converting them into the corresponding
normal distribution problems.
a. Note that P(Y ≤ 200) = P(In(Y) ≤ ln(200)) = P(X ≤ 5.30). We transform
x = 5.30 in the usual way to get From the z table,
we get P(Z ≤ 0.25) = 0.5987. Therefore. P(Y ≤ 200) = P(X ≤ 5.30) =
P(Z ≤ 0.25) = 0.5987.
b. The 90th percentile is a value y such that P (Y < y) = 0.90. We first note
that P(Y < y) = 0.90 is equivalent to p (ln(Y) < ln(y)) = P(X < x) = 0.90
where x = ln(y). We look up the cumulative probability of 0.90 in the z
table to get z = 1. 28. We use the inverse transformation to derive x = μ +
zσ = 5 + 1.28(1.2) = 6.54. Finally, we compute y = ex = e6.54 = 692.29.
Therefore, the 90th percentile of the distribution is 692.29.
F IG U R E 6 . 3 0 Using Excel to com pute a lognorm al cum ulative probability for a given x
• Standard_dev is the standard deviation of the normal distribution. We enter 1.2.
• Cumulative is a logical value. When we enter the value 1 or TRUE, Excel returns a
cumulative probability.
Thus, P ( Y ≤ 200) = 0.5982. Note that the earlier calculations were slightly off due to
rounding.
In order to find the 90th percentile of Y using Excel (Exercise 6.11b), we
select Formulas > Insert Function > LOGNORM.INV from Excel’s menu. In the
LOGNORM.INV dialog box, we supply three arguments (see Figure 6.31):
• Probability is a cumulative probability associated with the normal distribution. We
enter 0.90.
• Mean is the mean of the normal distribution. We enter 5.
• Standard_dev is the standard deviation of the normal distribution. We enter 1.2.
Thus, the 90th percentile is 690.81. Again, the earlier calculations were slightly off due
to rounding.
EXERCISES 6.4
Concepts 44. Compute the mean and the variance of a lognormal variable
Y = ex where X is normally distributed w ith the following
4 1 . A ran d o m variable X is e x p o n e n tia lly d is trib u te d w ith a mean
mean and variance:
o f 0.1.
a. μ = 3, σ 2 = 2
a. What is the rate parameter A? What is the standard
b. μ = 5, σ 2 = 2
deviation o f X?
c. μ = 5, σ 2 = 3
b. Compute P(X > 0.20).
c. Compute P(0.10≤ X ≤ 0.20). 45. Let Y = ex, where X is normally distributed. Compute the
mean and the variance of X given the following information.
42. A random variable X is exponentially distributed w ith an
a. μ r = 14, σ2Y = 22
expected value of 25.
b. μ r = 20, σ2Y= 2 2
a. What is the rate parameter λ ? What is the standard
deviation o f X? C. μ r = 20, σ 2 Y =
120
a. What is the rate parameter A? What is the standard $10,000. Use this inform ation to answer th e following
b. Jack wants to take a five-minute break. He believes that a. Compute the mean and the standard deviation of X.
if he goes right after he has serviced a customer, he will b. What proportion of the people in the Bahamas have
lower the probability of someone showing up during his household income above the mean?
five-minute break. Is he right in this belief? c. What proportion of the people in the Bahamas have
c. What is the probability that a customer will show up in household income below $20,000?
less than five minutes? d. Compute the 75th percentile of the income distribution
d. What is the probability that nobody shows up for over in the Bahamas.
half an hour? 51. It is well documented that a typical washing machine
49. When crossing the Golden Gate Bridge, traveling into can last anywhere between 5 to 12 years. Let the life
San Francisco, all drivers must pay a toll. Suppose the of a washing machine be represented by a lognormal
amount of time drivers wait in line to pay the toll follows an variable, Y = ex where X is normally distributed.
exponential distribution with a probability density function Furthermore, let the mean and standard deviation of
of f(x) = 0 .2e-2x the life of a washing machine be 8 years and 4 years,
.a \
What is the mean waiting time that drivers face when respectively.
entering San Francisco via the Golden Gate Bridge? a. Compute the mean and the standard deviation of X.
b. What is the probability that a driver spends more than b. What proportion of the washing machines will last for
the average time to pay the toll? more than 10 years?
c. What is the probability that a driver spends more than c. What proportion of the washing machines will last for
10 minutes to pay the toll? less than 6 years?
d. What is the probability that a driver spends between 4 d. Compute the 90th percentile of the life of the washing
and 6 minutes to pay the toll? machines.
Professor Lang is a professor of Economics at Salem State University. She has been
teaching a course in Principles of Economics for over 25 years. Professor Lang has never
graded on a curve since she believes that relative grading may unduly penalize (benefit)
a good (poor) student in an unusually strong (weak) class. She always uses an absolute
scale for making grades, as shown in the two left columns of Table 6.4.
TABLE 6 . 4 Grading Scales w ith Absolute Grading versus Relative Grading
Absolute Grading Relative Grading
Grade Score Grade Probability
A 92 and above A 0.10
B 78 up to 92 B 0.35
C 64 up to 78 C 0.40
D 58 up to 64 D 0.10
F Below 58 F 0.05
A colleague of Professor Lang’s has convinced her to move to rela
tive grading, since it corrects for unanticipated problems. Professor Lang
decides to experiment with grading based on the relative scale as shown
in the two right columns of Table 6.4. Using this relative grading scheme,
the top 10% of students will get As, the next 35% Bs, and so on. Based on
her years of teaching experience, Professor Lang believes that the scores in
her course follow a normal distribution with a mean of 78.6 and a standard
deviation of 12.4.
Professor Lang wants to use the above information to:
1. Calculate probabilities based on the absolute scale. Compare these
probabilities to the relative scale.
2. Calculate the range of scores for various grades based on the relative
scale. Compare these ranges to the absolute scale.
3. Determine which grading scale makes it harder to get higher grades.
Many teachers would confess that grading is one of the most difficult tasks of their profes Sam ple
sion. Two common grading systems used in higher education are relative and absolute. Rela
tive grading systems are norm referenced or curve based, in which a grade is based on the Report— ■
student’s relative position in class. Absolute grading systems, on the other hand, are criterion
referenced, in which a grade is related to the student’s absolute performance in class. In short, Absolute
with absolute grading, the student’s score is compared to a predetermined scale whereas with
relative grading, the score is compared to the scores of other students in the class. Grading versus
Let X represent the grade in Professor Lang’s class, which is normally distributed with
a mean of 78.6 and a standard deviation of 12.4. This information is used to derive the Relative
grade probabilities based on the absolute scale. For instance, the probability of receiving
an A is derived as P(X ≥ 92) = P(Z ≥ 1.08) = 0.14. Other probabilities, derived simi Grading
larly, are presented in Table 6.A.
TABLE 6.A Probabilities Based on Absolute Scale and Relative Scale
The second column of Table 6.A shows that 14% of students are expected to receive
As, 38% Bs, and so on. Although these numbers are generally consistent with the sug
gested relative scale shown in the third column of Table 6. A, it appears that the suggested
relative scale makes it harder for students to get higher grades. For instance, 14% get As
with the absolute scale compared to only 10% with the suggested relative scale.
Alternatively, we can compare the two grading methods on the basis of the range of
scores for various grades. The second column of Table 6.B restates the range of scores
based on absolute grading. In order to obtain the range of scores based on relative grading,
it is once again necessary to apply concepts from the normal distribution. For instance,
the minimum score required to earn an A with relative grading is derived by solving for a:
in P(X ≥ x) = 0.10. Since P(X ≥ x) = 0.10 is equivalent to P(Z ≥ z) = 0.10, it follows
that z = 1.28. Inserting the proper values of the mean, the standard deviation, and Z into
x = μ + z σ yields a value of x equal to 94.47. Ranges for other grades, derived similarly,
are presented in the third column of Table 6.B.
TABLE 6.B Range of Scores w ith Absolute Grading versus Relative Grading
Range of Scores Based on Range of Scores Based on
Grade Absolute Grading Relative Grading
A 92 and above 94.47 and above
B 78 up to 92 80.21 up to 94.47
C 64 up to 78 65.70 up to 80.21
D 58 up to 64 58.20 up to 65.70
F Below 58 Below 58.20
Once again comparing the results in Table 6.B, the use of the relative scale makes it harder
for students to get higher grades in Professor Lang’s courses. For instance, in order to receive
an A with relative grading, a student must have a score of at least 94.47 versus a score of at
least 92 with absolute grading. Both absolute and relative grading methods have their merits
and teachers often make the decision on the basis of their teaching philosophy. However, if
Professor Lang wants to keep the grades consistent with her earlier absolute scale, she should
base her relative scale on the probabilities computed in the second column of Table 6. A.
Conceptual Review
LO 6.1 Describe a continuous random variable.
A continuous random variable is characterized by (infinitely) uncountable val
ues and can take on any value within an interval. The probability that a continuous
random variable X assumes a particular value x is zero, that is, P(X = x) = 0. Thus,
for a continuous random variable, we calculate the probability within a specified
interval. Moreover, the following equalities hold: P(a ≤ X ≤ b) = P(a < X < b) =
P ( a ≤ X < b ) = P(a < X ≤ b).
The probability density function f(x) of a continuous random variable X describes
the relative likelihood that X assumes a value within a given interval. The probability
P(a ≤ X ≤ b) is the area under f(x ) between points a and b.
For any value x of the random variable X, the cumulative distribution function F(x)
is defined as F(x) = P(X ≤ x).
LO 6.5 Calculate and interpret probabilities for a random variable that follows the
normal distribution.
Any normally distributed random variable X with mean μ and standard deviation σ can
be transformed into the standard normal random variable Z as This normal
transformation implies that any value x of X has a corresponding value z of Z given by
A standard normal variable Z can be transformed to the normally distributed random vari
able X with mean μ and standard deviation σ as X = μ + Zσ . This inverse transformation
implies that any value z of Z has a corresponding value x of X given by x = μ + zσ .
LO 6.6 Calculate and interpret probabilities for a random variable that follows the
exponential distribution.
A useful nonsymmetric continuous probability distribution is the exponential d istri
bution. A random variable X follows the exponential distribution if its probability
density function is f(x) = λ e -λx for i ≥ 0, where A is a rate parameter and e ≈ 2 .7 18 is
the base of the natural logarithm. The mean and the standard deviation of the distribu
tion are both equal to 1/λ. For a given A, the cumulative probabilities are computed as
P(X < x) = 1 - e -λx.
LO 6.7 Calculate and interpret probabilities for a random variable that follows the
lognormal distribution.
The lognorm al distribution is another useful positively skewed distribution. Let X be a
normal random variable with mean μ and variance σ2 and let Y = ex be the correspond
ing lognormal variable. The mean μ r and standard deviation σ Y of Y are derived as
respectively. Equiva
lently, the mean and standard deviation of the normal variable X = ln(Y) are derived
respectively. Probabilities for a lognormal
55. U.S. consumers are increasingly viewing debit cards as a b. What is the probability o f a w aiting tim e less than
convenient substitute for cash and checks. The average amount 10 minutes between successive speeders?
spent annually on a debit card is $7,790 (Kiplinger's, August c. What is the probability of a w aiting tim e in excess o f
2007). Assume that the average amount spent on a debit card 25 minutes between successive speeders?
is normally distributed with a standard deviation of $500. 59. According to the Federal Bureau o f Investigation, there
a. A consumer advocate comments tha t the m ajority of is a violent crime in the U.S. every 22 seconds (ABCNews,
consumers spend over $8,000 on a debit card. Find a September 2 5 , 2007). Assume tha t the tim e between violent
flaw in this statement. crimes is exponentially distributed.
b. Compute the 25th percentile o f the am ount spent on a a. What is the probability tha t there is a violent crime in the
debit card. U.S. in the next one minute?
c. Compute the 75th percentile o f the am ount spent on a b. If there has not been a violent crime in the previous
debit card. minute, what is the probability tha t there w ill be a
d. W hat is the interquartile range o f this distribution? violent crime in the subsequent minute?
56. On St. Patrick's Day, men spend an average o f $43.87 while 60. The relief tim e provided by a standard dose o f a popular
wom en spend an average o f $29.54 (USA Today, March 17, children's allergy medicine averages six hours w ith a
2009). Assume the standard deviations o f spending for men standard deviation o f tw o hours.
and wom en are $3 and $11, respectively, and tha t both a. Determine the percentage o f children w ho experience
distributions are norm ally distributed. relief for less than four hours if the relief tim e follows a
a. W hat is the probability tha t men spend over $50 on normal distribution.
St. Patrick's Day? b. Determine the percentage o f children w ho experience
b. W hat is the probability th a t women spend over $50 on relief for less than four hours if the relief tim e follows a
St. Patrick's Day? lognormal distribution.
c. Are men or wom en more likely to spend over $50 on c. Compare the results based on these tw o distributions.
St. Patrick's Day? 61. The mileage (in thousands o f miles) that car owners get with
57. Lisa Mendes and Brad Lee work in the sales departm ent o f an a certain kind o f radial tire is a random variable Y having a
AT&T Wireless Store. Lisa has been signing in an average o f lognormal distribution such that Y = ex where X is normally
48 new cell phone customers every m onth w ith a standard distributed. Let the mean and the standard deviation of
deviation o f 22, while Brad signs in an average o f 56 new the life o f a radial tire be 40,000 miles and 5,000 miles,
customers w ith a standard deviation o f 17. The store manager respectively.
offers both Lisa and Brad a $100 incentive bonus if they can a. Compute the mean and standard deviation o f X.
sign in more than 100 new customers in a month. Assume a b. W hat proportion o f the tires w ill last for more than
normal distribution to answer the follow ing questions. 50.000 miles?
a. W hat is the probability th a t Lisa w ill earn the $100 c. W hat proportion o f the tires w ill last for no more than
incentive bonus? 35.000 miles?
b. W hat is the probability tha t Brad w ill earn the $100 d. Compute the 95th percentile o f the life distribution o f
incentive bonus? the tire.
c. Are you surprised by the results? Explain.
CASE STUDIES
Health officials of a midwestern town are concerned about the weight of children in their
town. For example, they believe that the BMI of their 10-year-old boys is normally dis
tributed with mean 19.2 and standard deviation 2.6.
In a report, use the sample information to:
1. Compute the proportion of 10-year-old boys in this town that are in the various
weight status categories given the BMI ranges.
2. Discuss whether the concern of health officials is justified.
LEARNING OBJECTIVES
After reading LO 7.1 Differentiate between a population parameter and a sample statistic.
this chapter
you should LO 7.2 Explain common sample biases.
be able to:
LO 7.3 Describe simple random sampling.
LO 7.5 Describe the properties of the sampling distribution of the sample mean.
LO 7.7 Describe the properties of the sampling distribution of the sample proportion.
LO 7.9 Construct and interpret control charts for quantitative and qualitative data.
In the last few chapters w e were given the population parameters, such as the population mean and
the population proportion, for the analysis o f discrete and continuous random variables. In many
instances we do not have information on the parameters, so we make inferences on the basis o f
sample statistics. Although sample statistics represent only a portion o f the population, they contain
useful inform ation to estimate the unknow n characteristics o f the population. The credibility o f any
statistical inference depends on the quality o f the sample on which it is based. In this chapter we
discuss various ways to draw a good sample and also h ighlight cases in which the sample misrepre
sents the population. It is im portant to note that any given situation involves only one population,
but many possible samples from which a statistic can be derived. Therefore, w hile the population pa
rameter is a constant, the sample statistic is a random variable whose value depends on the choice o f
the random sample. We will discuss h o w to evaluate the properties o f sample statistics. In particular,
we will study the probability distributions o f the sample mean and the sample proportion based on
simple random sampling. Finally, we will use these distributions to construct control charts, which
are popular statistical tools for m onitoring and im proving quality.
I N T R O D U C T O R Y CASE
P O P U L A T I O N VERSUS SAMPLE
A population consists of all items of interest in a statistical problem, whereas a
sample is a subset of the population. We use a calculated sample statistic, or sim
ply statistic, to make inferences about the unknown population parameter.
In later chapters we explore estimation and hypothesis testing, which are based on sample
information. It is important to note that no matter how sophisticated the statistical meth
ods are, the credibility of statistical inference depends on the quality of the sample on
which it is based. A primary requisite for a “good” sample is that it be representative of
the population we are trying to describe. When the information from a sample is not typi
cal of information in the population in a systematic way, we say that bias has occurred.
In addition to selection bias, the L iterary Digest survey also had a great deal of non
response bias. This occurs when those responding to a survey or poll differ systematically
from the nonrespondents. In the survey, a larger percentage of educated people mailed
back the questionnaires. During that time period, the more educated tended to come from
affluent families that again favored the Republican candidate. Problems with nonresponse
bias persist today. Most people do not want to spend time carefully reading and respond
ing to polls conducted by mail. Only those who care a great deal about an election or a
particular issue take the time to read the instructions, fill out the questionnaire, and mail it
back. Those who do respond may be atypical of the population as a whole.
What should the Literary Digest have done differently? At a minimum, most would agree
that names should have been obtained from voter registration lists rather than telephone
directory lists and car registrations. However, generating a sample by randomly selecting
names from voter registration lists also has possible shortcomings, especially with respect
to selection bias.
S IM P LE R A N D O M S A M P L E
A simple random sample is a sample of n observations which has the same prob
ability of being selected from the population as any other sample of n observations.
Most statistical methods presume simple random samples.
EXAMPLE 7.1
A recent analysis shows a dramatic decline in studying time among today’s college
students ( The Boston Globe, July 4, 2010). In 1961, students invested 24 hours
per week in their academic pursuits, whereas today’s students study an average of
14 hours per week. A dean at a large university in California wonders if this trend
is reflective of the students at her university. The university has 20,000 students and
the dean would like a sample of 100. Use Excel to draw a simple random sample of
100 students.
SOLUTION: Excel offers a useful tool to aid in randomly selecting 100 students
from the list of 20,000 students. Simply choose Formulas > Insert function >
RANDBETWEEN and input the number 1 for Bottom and the number 20,000 for
Top, as shown in Figure 7.1.
FIGURE 7.1 The dialog box for Excel's RANDBETWEEN fun ction
The RANDBETW EEN function will return a random number between these lim
its, perhaps the value 6,319. The dean can then choose the 6,319th student from
the list. In order to generate the remaining 99 random numbers, we can select the
cell with the value 6,319, drag it down 99 cells, and then from the menu choose
Home > Fill >
Down.
LO 7.4 While a simple random sample is the most commonly used sampling method, in some
Distinguish between situations other sampling methods have an advantage over simple random samples. Two
stratified random alternative methods for forming a sample are stratified random sampling and cluster
sampling and sampling.
cluster sampling. Political pollsters often employ stratified random sampling in an attempt to ensure
that each area of the country, each ethnic group, each religious group, and so forth, is ap
propriately represented in the sample. With stratified random sampling, the population is
divided into groups (strata) based on one or more classification criteria. Simple random
samples are then drawn from each stratum in sizes proportional to the relative size of each
stratum in the population. These samples are then pooled.
S T R A T IF IE D R A N D O M S A M P L IN G
In stratified random sampling, the population is first divided up into mutually exclu
sive and collectively exhaustive groups, called strata. A stratified sample includes
randomly selected observations from each stratum, which are proportional to the
stratum’s size.
Stratified random sampling has two advantages. First, it guarantees that the population
subdivisions of interest are represented in the sample. Second, the estimates of parameters
produced from stratified random sampling have greater precision than estimates obtained
from simple random sampling.
Even stratified random sampling, however, can fall short with its predictive ability.
One of the nagging mysteries of the 2008 Democratic presidential primaries was: Why
were the polls so wrong in New Hampshire? All nine major polling groups predicted that
Barack Obama would beat Hillary Clinton in the New Hampshire primary by an average
of 8.3 percentage points. When the votes were counted, Clinton won by 2.6%. Several
factors contributed to the wrong prediction by the polling industry. First, pollsters over
estimated the turnout of young voters, who overwhelmingly favored Obama in exit polls
but did not surge to vote as they had in the Iowa caucus. Second, Clinton’s campaign
made a decision to target women Democrats, especially single women. This focus did
not pay off in Iowa, but it did in New Hampshire. Finally, on the eve of the primary, a
woman in Portsmouth asked Clinton: “How do you do it?” Clinton’s teary response was
powerful and warm. Voters, who rarely saw Clinton in such an emotional moment, found
her response humanizing and appealing. Most polls had stopped phoning voters over the
weekend, too soon to catch the likely voter shift.
Cluster sampling is another method for forming a representative sample. A cluster
sample is formed by dividing the population into groups (clusters), such as geographic
areas, and then selecting a sample of the groups for the analysis. The technique works
best when most of the variation in the population is within the groups and not between the
groups. In such instances, a cluster is a miniversion of the population.
CLUSTER S A M P L IN G
In cluster sampling, the population is first divided up into mutually exclusive and
collectively exhaustive groups, called clusters. A cluster sample includes observa
tions from randomly selected clusters.
S T R A T IF IE D VERSUS CLUSTER S A M P L IN G
In stratified sampling, the sample consists of elements from each group, whereas in
cluster sampling, the sample consists of elements from the selected groups. Strati
fied sampling is preferred when the objective is to increase precision and cluster
sampling is preferred when the objective is to reduce costs.
E X E R C I S E S 7.1
1. In 2010, Apple introduced the iPad, a tablet-style com puter 40 students from her accounting class for the analysis.
tha t its form er CEO Steve Jobs called a "a truly magical and Discuss in detail whether or not Natalie's analysis is based on
revolutionary product'' (CNN, January 2 8 , 2010). Suppose you a representative sample.
are put in charge of determ ining the age profile o f people 4. Vons, a large supermarket in Grover Beach, California, is
who purchased the iPad in the U.S. Explain in detail the considering extending its store hours from 7:00 am to
follow ing sampling strategies th a t you could use to select a m idnight, seven days a week, to 6:00 am to m idnight. Discuss
representative sample. the sampling bias in the follow ing sampling strategies:
a. Simple random sampling a. Mail a prepaid envelope to randomly selected residents
b. Stratified random sampling in the Grover Beach area, asking for their preference for
c. Cluster sampling the store hours.
2. A m arke tin g firm opens a sm all b o o th a t a local m all over th e b. Ask the customers w ho frequent the store in the
w eekend, w here shoppers are asked h o w m uch m oney th e y morning if they would prefer an earlier opening time.
spent at th e fo o d court. The objective is to de term in e the c. Place an ad in the local newspaper, requesting people to
average m o n th ly expenditure o f shoppers at th e fo o d court. subm it their preference fo r store hours on the store's website.
Has th e m arketing firm c o m m itte d any sam pling bias? Discuss. 5. In the previous question regarding Vons' store hours, explain
3. Natalie Min is a business student in the Haas School of how you can obtain a representative sample based on the
Business at Berkeley. She wishes to pursue an MBA from follow ing sampling strategies:
Berkeley and wants to know the profile of other students a. Simple random sampling.
who are likely to apply to the Berkeley MBA program. In b. Stratified random sampling.
particular, she wants to know the GPA o f students w ith
c. Cluster sampling.
whom she m ight be competing. She randomly surveys
7.2 The Sampling Distribution
of the Sample Mean
As mentioned earlier, we are generally interested in the characteristics of the population.
For instance, a student is interested in the average starting salary (population mean) of
college graduates. Similarly, a banker is interested in the default probability (popula
tion proportion) of mortgage holders. Recall that the population mean and the population
proportion are the parameters that describe quantitative and qualitative data, respectively.
Since it is cumbersome or impossible to analyze the entire population, we generally make
inferences about the characteristics of the population on the basis of a random sample
drawn from the population.
It is important to note that there is only one population, but many possible samples
of a given size can be drawn from the population. Therefore, a population parameter is a
constant, even though its value may be unknown. A statistic, on the other hand, is a ran
dom variable whose value depends on the particular sample that is randomly drawn from
the population.
Consider the starting salary of a business graduate from a Big Ten school as the variable
of interest. If you decide to make inferences about the population mean on the basis of a
random draw of 38 recent graduates from a Big Ten school, then the sample mean is the
relevant statistic. However, the value of will change if you choose a different random
sample of 38 Big Ten students. The sample mean is commonly referred to as the estimator
of the population mean.
ESTIMATOR A N D ESTIMATE
When a statistic is used to estimate a parameter, it is referred to as an estimator.
A particular value of the estimator is called an estimate.
We continue to follow the convention introduced in earlier chapters where random vari
ables are denoted by upper-case letters and particular values of the random variables are
denoted by the corresponding lower-case letters. In the above example, the sample mean
is the estimator of the mean starting salary of Big Ten students. If the average derived
from a specific sample is $54,000, then s the estimate of the population mean.
Similarly, if the variable of interest is the default probability of mortgage holders, then
the sample proportion of defaults, denoted by from a random sample of 80 mortgage
holders is the estimator of the population proportion. If 10 out of 80 mortgage holders
in a given sample default, then is the estimate of the population
proportion.
In this section we will focus on the probability distribution of the sample mean
which is commonly referred to as the sampling distribution of Since is a random
variable, its sampling distribution is simply the probability distribution derived from all
possible samples of a given size from the population. Consider, for example, a mean
derived from a sample of n observations. Another mean can similarly be derived from a
different sample of n observations. If we repeat this process a very large number of times,
then the frequency distribution of the sample means can be thought of as its sampling
distribution. In particular, we will discuss the expected value and the standard deviation of
the sample mean. We will also study the conditions under which the sampling distribution
of the sample mean is normally distributed.
LO 7.5 The Expected Value and the Standard Deviation
Describe the of the Sample Mean
properties of the
Let the random variable X represent a certain characteristic of a population under study.
sampling distribution
Here X is a selected element of the population. Furthermore, for convenience, we refer
of the sample mean.
to X as the population of interest with an expected value, E ( X ) = μ . and a variance,
Var(X) = σ 2. Let the sample mean be based on a random sample of n observations
from the population. It is easy to derive the expected value and the variance of (see
Appendix 7.1 at the end of the chapter for the derivations).
The expected value of is the same as the expected value or the individual observa
tion, or E ( X ) = E (X ) = μ . In other words, if we were to sample repeatedly from a given
population, the average value of the sample means will equal the population mean from
the underlying population. This result holds irrespective of whether the sample mean is
based on a small or a large sample. We will elaborate on this desirable property of an
estimator in Chapter 8 on estimation.
The variance of is smaller than the variance of the individual obser
vation Var(X) = σ 2. This is an intuitive result, suggesting that averages have less varia
tion than individual observations. Since each sample is likely to contain both high and
low observations, the highs and lows cancel one another, making the variability between
sample means smaller than the variability between individual observations. As usual, the
standard deviation of the sampling distribution of is calculated as the positive square
root of the variance, or We often have to estimate this standard
deviation. The estimate is called the standard error of referred to as
EXAMPLE 7.2
The chefs at a local pizza chain in Cambria, California, strive to maintain the sug
gested size of their 16-inch pizzas. Despite their best efforts, they are unable to
make every pizza exactly 16 inches in diameter. The manager has determined that
the size of the pizzas is normally distributed with a mean of 16 inches and a standard
deviation of 0.8 inch.
a. What is the expected value and the standard deviation of the sample mean
derived from a random sample of 2 pizzas?
b. What is the expected value and the standard deviation of the sample mean
derived from a random sample of 4 pizzas?
c. Compare the expected value and the standard deviation of the sample mean
with those of an individual pizza.
SOLUTION: We know that the population mean μ = 16 and the population standard
deviation σ = 0.8. We use to calculate the following results.
a. With the sample size n = 2,
c. The expected value of the sample mean for both sample sizes is identical to
the expected value of the individual pizza. However, the standard deviation of
the sample mean with n = 4 is lower than the one with n = 2. For both sample
sizes, the standard deviation of the sample mean is lower than that of the
individual pizza. This result confirms that averaging reduces variability.
Sampling from a Normal Population
An important feature of the sampling distribution of the sample mean is that, irrespec
tive of the sample size n, is normally distributed if the population X from which the sam
ple is drawn is normal. In other words, if X is normal with expected value μ and standard
deviation σ , then is also normal with expected value μ and standard deviation
S A M P L IN G F R O M A N O R M A L P O P U L A T IO N
For any sample size n , the sampling distribution of is normal if the population X
from which the sample is drawn is normally distributed.
If is normal we can transform it into the standard normal random variable as:
EXAMPLE 7.3
Use the information in Example 7.2 to answer the following questions:
a. What is the probability that a randomly selected pizza is less than 15.5 inches?
b. What is the probability that 2 randomly selected pizzas average less than
15.5 inches?
c. What is the probability that 4 randomly selected pizzas average less than
15.5 inches?
d. Comment on the computed probabilities.
Note that when the sample size n = 1, the sample mean is the same as the
individual observation x.
a. In order to find P(X <
15.5), we transform x =
15.5 into
Therefore, P (X < 15.5) = P (Z < -0 .6 3 ) = 0.2643. There is a 26.43% chance
that an individual pizza is less than 15.5 inches.
b. Here we find We transform into
Therefore, In a random sample of
2 pizzas, there is an 18.94% chance that the average size is less than 15.5 inches.
c. Again we find but now n = 4. We transform into
Therefore In
a random sample of 4 pizzas, there is a 10.56% chance that the average size is
less than 15.5 inches.
d. The probability that the average size is under 15.5 inches, for 4 randomly
selected pizzas, is less than half of that for an individual pizza. This is due to
the fact that while X and have the same expected value of 16, the variance of
is less than that of X .
For any population X with expected value μ and standard deviation σ , the sampling
distribution of will be approximately normal if the sample size n is sufficiently
large. As a general guideline, the normal distribution approximation is justified
when n ≥ 30.
As before, if is approximately normal, then we can transform it to
Figure 7.3, discussed in Example 7.3, is not representative of the CLT principle because
for a normal population, the sampling distribution of is normal irrespective of the sam
ple size. Figures 7.4 and 7.5, however, illustrate the CLT by using random samples of
various sizes drawn from nonnormal populations. The relative frequency polygon of
Note that when the sample size n = 1, the sample mean is the same as the individual
observation (population) with the familiar uniform and exponential shapes. With n = 5,
the sampling distribution of begins to resemble the shape of the normal distribution.
With n = 30, the shapes of the sampling distribution of are approximately normal with
the uniform as well as the exponential population distributions. The CLT can similarly be
illustrated with other distributions of the population. How large a sample is necessary for
normal convergence depends on the magnitude of the departure of the population from
normality. As mentioned above, practitioners often use the normal distribution approxi
mation when n ≥ 30.
EXAMPLE 7.4
Consider the information presented in the introductory case of this chapter. Recall
that Anne wants to determine if the marketing campaign has had a lingering effect
on the amount of money customers spend on iced coffee. Before the campaign,
customers spent an average of $4.18 on iced coffee with a standard deviation of
$0.84. Anne reports that the average amount, based on 50 customers sampled after
the campaign, is $4.26. She wants to calculate the probability that customers spend
an average of $4.26 or more on iced coffee.
SOLUTION: If Starbucks did not pursue the marketing campaign, spending on iced
coffee would still have mean μ = 4.18 and standard deviation σ = 0.84. Anne
needs to calculate the probability that the sample mean is at least 4.26, that is,
The population from which the sample is drawn is not known to be
normal. However, since n ≥ 30, from the central limit theorem, we know that is
approximately normal. Therefore, we transform into
As shown in Figure 7.6,
It is quite plausible (probability = 0.2514) that in a sample o f 50
custom ers, the sample mean is $4.26 or more even if Starbucks did not pursue
the m arketing campaign.
FIGURE 7.6 Finding
E X E R C I S E S 7. 2
Concepts Applications
6. A random sample is drawn from a normally distributed 10. According to a recent survey, high school girls average 100 text
population w ith mean μ = 12 and standard deviation σ = 1.5. messages daily (The Boston Globe, April 2 1, 2010). Assume the
a. Comment on the sampling distribution of the sample population standard deviation is 20 text messages. Suppose a
mean w ith n = 20 and n = 40. random sample of 50 high school girls is taken.
b. Can you use the standard normal distribution to a. What is the probability that the sample mean is more
calculate the probability tha t the sample mean is less than 105?
than 12.5 for both sample sizes? b. W hat is the probability tha t the sample mean is less
c. Report the probability if you answered yes to the than 95?
previous question for either sample size. c. What is the probability tha t the sample mean is between
μ = 66 and standard deviation σ = 5.5. 11. Beer bottles are filled so tha t they contain 330 ml of beer in
a. Comment on the distribution o f the sample mean w ith each bottle. Suppose tha t the am ount o f beer in a bottle is
n = 16 and n = 36. normally distributed w ith a standard deviation of 4 ml.
b. Can you use the standard normal distribution to a. W hat is the probability tha t a random ly selected bottle
calculate the probability that the sample mean falls will have less than 325 ml of beer?
between 66 and 68 for both sample sizes? b. W hat is the probability tha t a randomly selected 6-pack
c. Report the probability if you answered yes to the o f beer will have a mean am ount less than 325 ml?
previous question for either sample size. c. What is the probability that a randomly selected 12-pack
8. A random sample o f size n = 100 is taken from a population of beer will have a mean am ount less than 325 ml?
w ith mean μ = 80 and standard deviation σ = 14. d. Comment on the sample size and the corresponding
for the sampling distribution of the sample mean. 12. Despite its nutritional value, seafood is only a tiny part of
b. W hat is the probability tha t the sample mean falls the American diet, w ith the average American eating just
between 77 and 85? 16 pounds o f seafood per year. Janice and Nina both work
in the seafood industry and they decide to create their own
c. W hat is the probability tha t the sample mean is greater
random samples and docum ent the average seafood diet
than 84?
in their sample. Let the standard deviation of the American
9. A random sample of size n = 50 is taken from a population
seafood diet be 7 pounds.
w ith mean μ = - 9 . 5 and standard deviation σ = 2.
a. Janice samples 42 Americans and finds an average
a. Calculate the expected value and the standard deviation
seafood consumption of 18 pounds. How likely is it
for the sampling distribution of the sample mean.
to get an average o f 18 pounds or more if she had a
b. What is the probability that the sample mean is less than — 10?
representative sample?
c. W hat is the probability th a t the sample mean falls
b. Nina samples 90 Americans and finds an average
between - 1 0 and —9?
seafood consumption of 17.5 pounds. How likely is it
to get an average o f 17.5 pounds or more if she had a b. Find the probability tha t the mean w eight of
representative sample? 16 randomly selected turkeys is more than 25 pounds.
c. Which of the tw o women is likely to have used a more c. Find the probability tha t the mean w eight of 16
representative sample? Explain. randomly selected turkeys is between 18 and 24 pounds.
13. The w e ig h t o f people in a small to w n in Missouri is known 15. A small hair salon in Denver, Colorado, averages about 30
to be norm ally d istributed w ith a mean o f 180 pounds and a customers on weekdays w ith a standard deviation of 6. It is
standard deviation o f 28 pounds. On a raft th a t takes people safe to assume that the underlying distribution is normal. In
across th e river, a sign states, "M a xim u m capacity 3,200 pounds an attem pt to increase the number of weekday customers,
o r 16 persons." W hat is th e p ro b a b ility th a t a random sample o f the manager offers a $2 discount on 5 consecutive weekdays.
16 persons w ill exceed th e w e ig h t lim it o f 3,200 pounds? She reports that her strategy has worked since the sample
14. The weight o f turkeys is known to be normally distributed mean of customers during this 5 weekday period jum ps to 35.
w ith a mean of 22 pounds and a standard deviation of a. How unusual would it be to get a sample average of 35 or
5 pounds. more customers if the manager had not offered the discount?
a. Discuss the sampling distribution of the sample mean b. Do you feel confident that the manager's discount
based on a random draw o f 16 turkeys. strategy has worked? Explain.
SOLUTION: Given that p = 0.77, we can derive the expected value and the standard
deviation as follows.
a. With n = 100,
b. With n = 200,
c. As in the case of the sample mean, while the expected value of the sample pro
portion is unaffected by the sample size, the standard deviation of the sample
proportion is reduced as the sample size increases.
According to the CLT, the sampling distribution of approaches the normal distribu
tion as the sample size increases. However, as the population proportion deviates from
p = 0.50, we need a larger sample size for the approximation. We illustrate these results
by generating the sampling distribution of from repeated draws from a population with
various values of the population proportion and sample sizes. As in the case of we use
the relative frequency polygon to represent the distribution of The simulated sampling
distribution of is based on the population proportion p = 0.10 (Figure 7.7) and p = 0.30
(Figure 7.8).
F IG U R E 7 .7 Sampling distributio n o f when the
population pro po rtion is p = 0.10
When p = 0.10, the sampling distribution of does not resemble the shape of the normal
distribution with n = 20 since the approximation condition np ≥ 5 and n (1 —p ) ≥ 5 is
not satisfied. However, the curve becomes close to normal with n = 100. When p = 0.30,
the shape of the sampling distribution of is approximately normal since the approxi
mation condition is satisfied with both sample sizes. In empirical work, it is common to
work with large survey data, and as a result, the normal distribution approximation is
justified.
EXAMPLE 7.6
Consider the information presented in the introductory case of this chapter. Recall
that Anne Jones wants to determine if the marketing campaign has had a lingering
effect on the proportion of customers who are women and teenage girls. Prior to the
campaign, 43% of the customers were women and 21% were teenage girls. Based
on a random sample of 50 customers after the campaign, these proportions increase
to 46% for women and 34% for teenage girls. Anne has the following questions.
a. If Starbucks chose not to pursue the marketing campaign, how likely is it that
46% or more of iced-coffee customers are women?
b. If Starbucks chose not to pursue the marketing campaign, how likely is it that
34% or more of iced-coffee customers are teenage girls?
SOLUTION: If Starbucks had not pursued the marketing campaign, the proportion
of customers would still be p = 0.43 for women and p = 0.21 for teenage girls.
With n = 50, the normal approximation of the sample proportion is justified for
both population proportions.
a . In order to calculate we transform into
With only a 1.19% chance, it is unlikely that the proportion of iced coffee
purchased by teenage girls is at least 0.34 if Starbucks did not pursue the
marketing campaign.
Therefore, Anne can use this sample information to infer that the increase in the
proportion of iced-coffee sales to women may not necessarily be due to the market
ing campaign. However, the marketing campaign seems to have been successful in
increasing the proportion of iced-coffee sales to teenage girls.
S Y N O P S I S OF I N T R O D U C T O R Y C A S E
E X E R C I S E S 7. 3
Concepts Applications
16. Consider a population proportion p = 0.68. 19. A recent study by Allstate Insurance Co. finds that 82% of
a. Calculate the expected value and the standard deviation teenagers have used cell phones while driving (The Wall
of w ith n = 20. Is it appropriate to use the normal Street Journal, May 5 , 2010). Suppose a random sample of
distribution approxim ation for Explain. 100 teen drivers is taken.
b. Calculate th e e x p e c te d value a n d th e sta n d a rd deviation a. Discuss the sampling distribution of the sample
of w ith n = 50. Is it appropriate to use the normal proportion.
distribution approximation for Explain. b. What is the probability that the sample proportion is less
17. Consider a population proportion p = 0.12. than 0.80?
a. Discuss the sampling distribution o f the sample c. What is the probability tha t the sample proportion is
proportion w ith n = 20 and n = 50. w ith in ± 0.02 o f the population proportion?
b. Can you use the normal approximation to calculate the 20. According to a recent FCC survey, one in six cell phone users
probability that the sample proportion is between 0.10 have experienced "bill shock" from unexpectedly high cell
and 0.12 for both sample sizes? phone bills (Tech Daily Dose, May 26, 2010).
c. Report the probabilities if you answered yes to the a. Discuss the sampling distribution of the sample
previous question. proportion based on a sample of 200 cell phone
users. Is it appropriate to use the normal distribution
18. A random sample o f size n = 200 is taken from a population
approximation for the sample proportion?
w ith population proportion p = 0.75.
b. What is the probability that more than 20% o f cell phone
a. Calculate the expected value and the standard deviation
users in the sample have experienced "bill shock"?
for the sampling distribution of the sample proportion.
b. What is the probability tha t the sample proportion is 21. A car manufacturer is concerned about poor customer
between 0.70 and 0.80? satisfaction at one o f its dealerships. The management
decides to evaluate the satisfaction surveys of its next
c. What is the probability that the sample proportion is less
40 customers. The dealer w ill be fined if fewer than
than 0.70?
26 customers report favorably and the dealership will be a m achine. The m achine is know n to have 60 green balls
dissolved if fewer than 22 report favorably. It is known and 40 red balls. After each draw, the color o f th e ball is
that 70% of the dealer's customers report favorably on noted and the ball is pu t back fo r the next draw. You w in
satisfaction surveys. a prize if m ore than 70% o f the draws result in a green
a. What is the probability that the dealer will be fined? ball. W ould you choose 50 or 100 draws fo r the game?
22. Europeans are increasingly upset at their leaders for 24. After years of rapid growth, illegal im m igration into the
making deep budget cuts to many social programs that United States has declined, perhaps owing to the recession
are becoming too expensive to sustain. For instance, the and increased border enforcement by the U.S. (Los Angeles
popularity o f President Nicolas Sarkozy o f France has Times, September 1 , 2010). While its share has declined,
plummeted, giving him an approval rating o f just 26% (The California still accounts for 23% of the nation's estimated
a. What is the probability tha t fewer than 60 of 200 French a. In a sample of 50 illegal immigrants, what is the
people give President Sarkozy a favorable rating? probability that more than 20% live in California?
b. What is the probability that more than 150 of 200 French b. In a sample o f 200 illegal immigrants, w hat is the
people give President Sarkozy an unfavorable rating? probability tha t more than 20% live in California?
c. Comment on the reason for the difference between the
23. At a new e x h ib it in the Museum o f Science, people are
computed probabilities in parts a and b.
asked to choose betw een 50 or 100 random draws from
TH E F IN IT E P O P U L A T IO N C O R R E C T IO N
FACTOR FOR THE SA M PLE M E A N
We use the finite p o p u lation co rrectio n fa cto r to reduce the sampling variation of
The resulting standard deviation is The transformation of
to Z is made accordingly.
Note that the correction factor is always less than one; when N is large relative to n, the
correction factor is close to one and the difference between the formulas with and without
the correction is negligible.
EXAMPLE 7.7
A large introductory marketing class has 340 students. The class is divided up into
10 groups for the final course project. Connie is in a group of 34 students. These
students had averaged 72 on the midterm, when the class as a whole had an average
score of 73 with a standard deviation of 10. Connie is concerned that her group is
not very good.
a. Calculate the expected value and the standard deviation of the sample mean.
b. How likely is it that a random sample of 34 students will average 72 or lower?
SOLUTION: The population mean is μ = 73 and the population standard deviation
is σ = 10.
a. The expected value of the sample mean is We use the finite
population correction factor because the sample size n = 34 is more than
5% of the population size N = 340. Therefore, the standard deviation of
the sample mean is
Note that without the correction factor, the standard deviation would be
higher at
b. In order to calculate P (X ≤ 72), we use to transform into
Therefore,
That is, the likelihood of 34 students averaging 72 or lower is 27.09%.
We can use a similar finite population correction factor for a sample proportion when the
sample size is more than 5% of the population size.
THE F IN IT E P O P U L A T IO N C O R R E C T IO N
FACTO R FOR THE S A M P L E P R O P O R T IO N
We use the finite population correction factor to reduce the sampling varia
tion of the sample proportion The resulting standard deviation of is
The transformation of to Z is made accordingly.
EXAMPLE 7.8
The home ownership rate during 2009 declined to approximately 67% and is now
comparable to the rate in early 2000 ( U.S. Census Bureau News, February 2, 2010).
A random sample of 80 households is taken from a small island community with
1,000 households. The home ownership rate on the island is equivalent to the
national home ownership rate of 67%.
a. Calculate the expected value and the standard deviation for the sampling dis
tribution of the sample proportion. Is it necessary to apply the finite population
correction factor? Explain.
b. What is the probability that the sample proportion is within 0.02 of the popula
tion proportion?
SOLUTION:
a. We must apply the finite population correction factor because the sample
size n = 80 is more than 5% of the population size N = 1,000. Therefore,
and
b. The probability that the sample proportion is within 0.02 of the population pro
portion is We transform into
—0.40. Similarly, we transform into z = 0.40. Therefore, P (0.65 ≤
The likeli
hood of getting a home ownership rate within 0.02 of the population propor
tion is 31.08%.
E X E R C I S E S 7. 4
Concepts Exchange C o m m ittee (SEC) w ebsite, it w as d e te rm in e d th a t
th e m ean a n d th e sta n d ard deviation of c o m p e n sa tio n for
25. A ra n d o m sa m p le of size n = 100 is tak e n from a p o p u latio n
th e 500 h ig h est paid CEOs in publicly tra d e d U.S. c o m pan ies
of size N = 2,500 w ith m ea n μ = —45 a n d variance σ2 = 81.
is $10.32 million a n d $9.78 million, respectively. An analyst
a. Is it n e cessary to ap p ly th e finite p o p u latio n correction
random ly cho o ses 32 CEO c o m p e n sa tio n s for 2006.
factor? Explain. Calculate th e e x p e c te d value a n d th e
a. Is it n ecessary to a pp ly th e finite p o p u latio n correction
sta n d a rd dev iatio n of th e sa m p le m ean.
factor? Explain.
b. W hat is th e probability th a t th e sa m p le m ean is b e tw e e n
b. Is th e sam p lin g d istrib u tio n o f th e sa m p le m ea n
– 47 a n d – 43?
norm ally distrib u ted ? Explain.
c. W hat is th e probability th a t th e sa m p le m ean is g re a te r
c. Calculate th e e x p e c te d value a n d th e s ta n d a rd deviation
th a n – 44?
o f th e sa m p le m ean .
26. A random sam ple of size n = 70 is taken from a finite population
d. W h at is th e probability th a t th e sa m p le m ean is m ore
o f size N = 500 with m ean μ = 220 a n d variance σ 2 = 324.
th a n $12 million?
a. Is it n ecessary to apply th e finite p o p u latio n correction
30. S u p p o se in th e previous q u e stio n th e a n alyst h ad ran do m ly
factor? Explain. C alculate th e e x p e c te d v alue a n d th e
c h o se n 12 CEO c o m p e n sa tio n s for 2006.
sta n d a rd deviation of th e sa m p le m ean.
a. Is it n ecessary to a pp ly th e finite p o p u latio n correction
b. W hat is t h e pro bab ility t h a t th e sa m p le m ea n is less
factor? Explain.
th a n 210?
b. Is th e sam p lin g d istrib u tio n o f th e sa m p le m ean
c. W hat is th e probability th a t th e sa m p le m ea n lies
norm ally d istrib u ted ? Explain.
b e tw e e n 215 a n d 230?
c. C alculate th e e x p e c te d value a n d th e s ta n d a rd deviation
27. A ra n d o m sa m p le of size n = 100 is ta k e n from a p o p u latio n
of th e sa m p le m ean.
o f size N = 3,000 w ith a p o p u latio n p ro p o rtio n o f p = 0.34.
d. Can you u se th e norm al ap p ro x im atio n to calculate
a. Is it n ecessary to apply th e finite p o p u latio n correction
t h e probability t h a t th e sa m p le m ean is m o re th a n
factor? Explain. Calculate th e e x p e c te d v alue a n d th e
$12 m illion? Explain.
s ta n d a rd dev iatio n of th e sa m p le p ro p o rtio n .
31. Given th e recent econom ic dow nturn, only 60% in a g raduating
b. W hat is th e probability th a t th e sa m p le p ro p o rtio n is
class of 250 will find em p lo y m en t in th e first round of a jo b
g re a te r th a n 0.37?
search. You have 20 friends w h o have recently graduated.
28. A ra n d o m sa m p le of size n = 80 is ta k e n from a p o p u latio n
a. Discuss th e sam pling d istrib u tio n of th e sam p le
o f size N = 600 w ith a p o p u latio n p ro p o rtio n p = 0.46.
p ro p o rtio n o f your friends w h o will find e m p lo y m e n t in
a. Is it necessary to a p p ly th e finite p o p u latio n correction th e first ro un d o f a jo b search.
factor? Explain. Calculate th e e x p e c te d v alue a n d th e
b. W hat is th e probability th a t less th a n 50% of your friends
sta n d a rd dev iatio n o f th e sa m p le pro p o rtio n .
will find e m p lo y m e n t in th e first ro und of a jo b search?
b. W h at is th e prob ab ility th a t th e sa m p le m ea n is less
32. Despite th e recession, com panies are setting aside a large chunk
th a n 0.40?
of their IT spending for green technology projects (BusinessWeek,
March 5 , 2009). Two o u t of th ree of th e large com panies
Applications surveyed by Deloitte said th ey have at least 5% of their IT b u d g e t
29. The issues su rrou n din g th e levels a n d stru ctu re of executive earm arked for green IT projects. S uppose th a t th e survey was
c o m p e n sa tio n have gain ed a d d e d p ro m in en ce in th e w ake of based on 1,000 large com panies. W hat is th e probability th a t
th e financial crisis th a t e ru p te d in th e fall o f 2008. Based on m ore th an 75 of 120 large com panies will have at least 5% of
th e 2006 c o m p e n sa tio n d a ta o b tain e d from th e Securities and their IT expenditure earm arked for green IT projects?
In general, two approaches are used for statistical quality control. A firm uses acceptance
sampling if it produces a product (or offers a service) and at the completion of the pro
duction process, the firm then inspects a portion of the products. If a particular product
does not conform to certain specifications, then it is either discarded or repaired. The
problems with this approach to quality control are, first, it is costly to discard or repair a
product. Second, the detection of all defective products is not guaranteed. Defective prod
ucts may be delivered to customers, thus damaging the firm’s reputation.
A preferred approach to quality control is the detection approach. A firm using
the detection approach inspects the production process and determines at which point
the production process does not conform to specifications. The goal is to determine
whether the production process should be continued or adjusted before a large number
of defects are produced. In this section we focus on the detection approach to quality
control.
In general, no two products or services are identical. In any production process, varia
tion in the quality of the end product is inevitable. Two types of variation occur. Chance
variation, or common variation, is caused by a number of randomly occurring events
that are part of the production process. This type of variation is not generally considered
under the control of the individual worker or machine. For example, suppose a machine
fills one-gallon jugs of milk. It is unlikely that the filling weight of each jug is exactly
128 ounces. Very slight differences in the production process lead to minor differences
in the weights of one jug to the next. Chance variation is expected and is not a source of
alarm in the production process so long as its magnitude is tolerable and the end product
meets acceptable specifications.
The other source of variation is referred to as assignable variation, or specially caused
variation. This type of variation in the production process is caused by specific events or
factors that can usually be identified and eliminated. Suppose in the milk example that the
machine is “drifting” out of alignment. This causes the machine to overfill each jug— a
costly expense for the firm. This is the type of variation in the production process that the
firm wants to identify and correct.
The most commonly used statistical tool in quality control is the control chart, a
plot of calculated statistics of the production process over time. If the calculated
statistics fall in an expected range, then the production process is in control. If the
calculated statistics reveal an undesirable trend, then adjustment of the production
process is likely necessary.
We first focus on quantitative data and examine the chart, which monitors the central
tendency of a production process. We then move to qualitative data and construct the
chart, which assesses a process by showing the proportion of defectives (or some other
characteristic) in a series of samples. We use the following example to elaborate on the
chart.
EXAMPLE 7.9
A firm that produces one-gallon jugs of milk wants to ensure that the machine is
operating properly. Every two hours, the company samples 25 jugs and calculates
the following sample mean filling weights (in ounces):
Assume that when the machine is operating properly, μ = 128 and σ = 2, and
that filling weights follow the normal distribution. Can the firm conclude that the
machine is operating properly? Should the firm have any concerns with respect to
this machine?
In the given example, the upper control limit is and the lower
control limit is Figure 7.11 shows the centerline and the con
trol limits as well as the sample means for Example 7.9.
If the UCL equation provides a value greater than one, we reset UCL to one in the control
chart. Similarly, if the LCL equation provides a negative value for LCL, we reset LCL to
zero in the control chart.
EXAMPLE 7.10
In a 2010 automotive quality study conducted by J .D. Power & Associates,
Toyota Motor Corp. was rated below average, falling to 21st place from 7th the
year before (The Wall Street Jo urna l, June 18, 2010). Toyota’s reputation suffered
due to massive recalls in late 2009 and early 2010 related to defects that could
cause the com pany’s cars to accelerate on their own. Moreover, the company was
criticized for its delay in responding to the problem. Suppose that from past ex
perience, Toyota expects complaints from 5% of new car buyers each month. For
each of the last six months, the company surveys 500 new car buyers and asks if
the owner has any complaints about the car. The following sample proportions
are obtained:
SOLUTION:
a. Since historical data indicate that 5% of new car buyers lodge some type of
complaint, we set the centerline at p = 0.05. We then calculate the upper con
trol limit and lower control limit as follows:
Plotting the values for the centerline, UCL, and LCL, as well as the sample
proportions, yields Figure 7.13.
F IG U R E 7 .1 3 Proportion chart for com plaints
b. Four of the most recent sample proportions fall above the upper control limit.
The proportion of complaints far surpasses what Toyota expects from past
experience. Based on the given sample information, the company needs to
seriously address the complaints to avoid further harm to its reputation.
In order to construct a chart using Excel, you would follow the same steps as those
outlined earlier for the chart.
E X E R C I S E S 7. 5
13.7 14.2 13.9 14.1 14.3 13.9 0.044 0.052 0.060 0.036 0.028 0.042 0.034 0.054 0.048 0.025
a. Construct an chart. Plot the sample means on the a. Construct a chart. Plot the sample proportions on the
chart. chart.
b. Can the firm conclude tha t the production process is b. Can the firm conclude that the process is operating
operating properly? Explain. properly?
40. Major League Baseball Rule 1.09 states tha t "the baseball 43. A m anufacturing process produces steel rods in batches of
shall weigh not less than 5 or more than 5¼ ounces" 1,000. The firm believes tha t the percent o f defective items
(www.m lb.com ). Use these values as the lower and the generated by this process is 5%.
upper control limits, respectively. Assume the centerline a. Construct the centerline and the upper and lower control
equals 5.125 ounces. Periodic samples of 50 baseballs limits for th e chart.
produce the follow ing sample means:
b. An engineer inspects the next batch of 1,000 steel rods
5.05 5.10 5.15 5.20 5.22 5.24 and finds tha t 6.2% are defective. Is the manufacturing
process under control? Explain.
a. Construct an chart. Plot th e sam ple means on th e
44. The college admissions office at a local university usually
chart.
admits 750 students and knows from previous experience
b. Are any points outside the control limits? Does it appear
tha t 25% of these students choose not to enroll at the
that the process is under control? Explain. university. Assume p = 0.25.
41. FILE Fast bowling, also known as pace bowling, is an a. Construct the centerline and the upper and lower
im p o rta n t com ponent o f the bow ling attack in the sport of control limits for th e chart.
cricket. The objective is to bowl at high speed and make the b. Assume that this year the university admits
ball turn in the air and o ff the ground so th a t it becomes 750 students and 240 choose not to enroll at the
d ifficu lt for the batsman to hit it cleanly. Kalwant Singh university. Should the university be concerned?
is a budding Indian cricketer in a special bow ling camp.
Explain.
W hile his coach is happy w ith Kalwant's average bowling
45. Following customer com plaints about the quality of
speed, he feels th a t Kalwant lacks consistency. He records
service, Dell stopped routing corporate customers to
his bow ling speed on the next fou r overs, where each
a technical support call center in Bangalore, India
over consists of six balls; the data labeled Cricket is also
(USA Today, November 2 4 , 2003). Suppose Dell's decision to
available on the text website.
direct customers to call centers outside o f India was based
Over 1 Over 2 Over 3 Over 4 on consumer complaints in the last six months. Let the
96.8 99.2 88.4 98.4 num ber o f complaints per m onth for 80 randomly selected
99.5 100.2 97.8 91.4 customers be given below.
Barbara Dwyer, the manager at Lux Hotel, makes every effort to ensure
that customers attempting to make phone reservations wait an average of
only 60 seconds to speak with a reservations specialist. She knows that this
is likely to be the customer’s first impression of the hotel and she wants the
initial interaction to be a positive one. Since the hotel accepts phone reser
vations 24 hours a day, Barbara wonders if this quality service is consis
tently maintained throughout the day. She takes six samples of n = 4 calls
during each of four shifts over one 24-hour period and records the wait time
of each call. A portion of the data, in seconds, is presented in Table 7.1; the
complete data, labeled Lux Hotel , are available on the text website.
Barbara relays to the quality control engineer that the mean wait time is 60 seconds with
a standard deviation of 30 seconds. She instructs the quality control engineer to use the
sample information to:
1. Prepare a control chart for the wait time for a customer.
2. Use the control chart to determine whether the quality service is consistently main
tained throughout the day.
When a potential customer phones Lux Hotel, it is imperative for the reservations special Sample
ist to set a tone that relays the high standard of service that the customer will receive if he/
she chooses to stay at the Lux. For this reason, management at the Lux strives to minimize Report—
the time that elapses before a potential customer speaks with a reservations specialist;
however, management also recognizes the need to use its resources wisely. If too many Customer
reservations specialists are on duty, then resources are wasted due to idle time; yet if too
few reservations specialists are on duty, the result might mean angry first-time customers, Wait Time
or worse, lost customers. In order to ensure customer satisfaction as well as an efficient use
of resources, a study is conducted to determine whether a typical customer waits an aver
age of 60 seconds to speak with a reservations specialist. Before data are collected, a con
trol chart is constructed. The upper control limit (UCL) and the lower control limit (LCL)
are set three standard deviations from the desired average of 60 seconds. In Figure 7.A,
the desired average of 60 seconds is denoted as the centerline and the upper and lower
control limits amount to 105 seconds and 15 seconds
respectively. The reservation process is deemed under control if the sample means fall
randomly within the upper and lower control limits; otherwise the process is out of con
trol and adjustments should be made.
During each of four shifts, six samples of n = 4 calls are randomly selected over one
24-hour period and the average wait time of each sample is recorded. All six sample
means from the first shift ( 1st shift: 12:00 am -6:00 am, sample numbers one through
six) fall within the control limits, indicating that the reservation process is in control.
However, five sample means from the second shift (2nd shift: 6:00 am -12:00 pm, sample
numbers seven through 12) lie above the upper control limit. Customers calling during the
second shift are waiting too long before they speak with a specialist. In terms of quality
standards, this is unacceptable from the hotel’s perspective. All six sample means from
the third shift fall within the control limits (3rd shift: 12:00 pm -6:00 pm, sample numbers
13 through 18), yet all sample means for the fourth shift fall below the lower control limit
(4th shift: 6:00 pm -12:00 am, sample numbers 19 through 24). Customers are waiting
for very short periods of time to speak with a reservations specialist, but reservations spe
cialists may have too much idle time. Perhaps one solution is to shift some reservations
specialists from shift four to shift two.
Conceptual Review
LO 7.1 D ifferentiate betw een a population param eter and a sample statistic.
The numerical measure of a population is called a population parameter, or simply
a parameter. A parameter is a constant even though its value may be unknown. The
numerical measure of a sample is called a sample statistic, or simply a statistic. A sta
tistic, such as the sample mean or the sample proportion, is a random variable whose
value depends on the chosen random sample.
LO 7.5 Describe the properties of the sampling distribution of the sample mean.
Since the sample mean is a random variable, its sampling distribution is the probability
distribution of sample means derived from all possible samples of a given size from
the population. The expected value of the sample mean equals and the
standard deviation equals For any sample size, the sampling distribution
of is normal if the population is normally distributed.
A normally distributed can be transformed into a standard normal random variable
as Therefore, any value on has a corresponding value z on Z given by
LO 7.7 Describe the properties of the sampling distribution of the sample proportion.
The expected value and the standard deviation of the sample proportion are
The central limit theorem states that for any popula
tion proportion p, the sampling distribution of is approximately normal if the sample
size n is sufficiently large. As a general guideline, the normal distribution approxima
tion is justified when n ≥ 5 and n(1 – p ) ≥ 5. If we assume to be normal, it can be
transformed into a standard normal random variable as Therefore, any
b. Stratified random sampling. b. Find the probability tha t the average length o f 10
a. What is the probability tha t 8 workers over the age of selling products such as gourm et foods, beer and wine, bread,
55 take an average of more than 20 weeks to find a job? nuts, cereal, and coffee. One o f their best-selling nuts is Raw
California Almonds, which are priced at $4.49 for 16 ounces.
b. W hat is the probability tha t 20 younger workers average
Since it is impossible to pack exactly 16 ounces in each packet,
less than 15 weeks to find a job?
a researcher has determined that the w eight of almonds in
49. While starting salaries have fallen for college graduates
each packet is normally distributed w ith a mean and standard
in many o f the to p hiring fields, there is some good news
deviation equal to 16.01 and 0.08 ounces, respectively.
for business undergraduates w ith concentrations in
a. Discuss the sampling distribution o f the sample mean Street Journal public opinion poll, President Barack Obama
based on any given sample size. has reached his lowest approval rating since taking office
b. Find the probability that a random sample of 20 bags of in January o f 2009. The poll showed that 48% o f people
almonds will average less than 16 ounces. disapprove of the jo b Obama is doing as President of the
c. Suppose your cereal recipe calls for no less than United States, while only 45% approve. Experts attribute
48 ounces o f almonds. What is the probability that three the drop in approval ratings to a poor economy and the
packets of almonds will meet your requirement? government's reaction to the massive oil spill in the Gulf of
Mexico. Assume the above approval and disapproval ratings
52. The average college graduate has a student loan debt
in June 2010 to answer the following questions.
of $23,186 (The Boston Globe, June 8 , 2010). Assume the
a. What is the probability that President Obama gets a
population standard deviation is $5,000. Suppose you take a
majority support in a random sample of 50 Americans?
random sample of 50 college graduates.
b. What is the probability that President Obama gets
a. What is the probability tha t the sample mean debt is
a majority disapproval in a random sample of 50
more than $25,000?
Americans?
b. What is the probability that the sample mean debt is less
than $22,000? 57. The producer of a particular brand of soup claims that its
sodium content is 50% less than that of its competitor.
c. What is the probability that the sample mean debt is
The food label states that the sodium content measures
w ithin $2,500 o f the population mean?
410 milligrams per serving. Assume the population of
53. Data from the Bureau o f Labor Statistics'Consumer
sodium content is normally distributed w ith a standard
Expenditure Survey (CE) show th a t annual expenditures
deviation of 25 milligrams. Inspectors take periodic samples
for cellular phone services per consumer u n it increased
of 25 cans and measure the sodium content. The following
from $210 in 2001 to $608 in 2007. Let the standard
sample means are obtained.
deviation o f annual cellular expenditure be $48 in 2001
and $132 in 2007. 405 412 399 420 430 428
a. What is the probability that the average expenditure of a. Construct an chart. Plot th e sam ple m ea n s on th e
100 cellular customers in 2001 exceeded $200? chart.
b. What is the probability that the average expenditure of b. Can the inspectors conclude that the producer is
100 cellular customers in 2007 exceeded $600? advertising the sodium content accurately? Explain.
54. According to a recent report, scientists in New England say 58. FILE A variety of packaging solutions exist for products
they have identified a set o f genetic variants that predicts tha t must be kept w ith in a specific tem perature range.
extreme longevity with 77% accuracy (New York Times, Cold chain distribution is particularly useful in the food and
July 1 , 2010). Assume 150 patients decide to get their pharmaceutical industries. A packaging company strives
genome sequenced. to maintain a constant tem perature for its packages. It is
a. If the claim by scientists is accurate, what is the believed th a t the tem perature of its packages follows a
probability that more than 120 patients will get a correct normal distribution w ith a mean of 5 degrees Celsius and
diagnosis for extreme longevity? a standard deviation of 0.3 degree Celsius. Inspectors take
b. If the claim by scientists is accurate, what is the weekly samples for 5 weeks of eight randomly selected
probability tha t fewer than 70% o f the patients will get a boxes and report the follow ing temperatures in degrees
correct diagnosis for extreme longevity? Celsius. A po rtion o f the data are given below; the complete
55. American workers are increasingly planning to delay data, labeled Packaging, are available on the text website.
retirement (US News & World Report, June 3 0 , 2010). Week 1 Week 2 Week 3 Week 4 Week 5
According to a Pew Research Center comprehensive survey, 3.98 5.52 5.79 3.98 5.14
35% of employed adults of age 62 and older say they have
4.99 5.52 6.42 5.79 6.25
pushed back their retirement date.
a. What is the probability that in a sample of 100 employed
4.95 4.95 5.44 5.95 4.28
adults of age 62 and older, more than 40% have pushed
back their retirement date? a. Construct an c h a rt for quality control. Plot th e five
b. What is the probability tha t in a sample of 200 employed w eekly sa m p le m ea n s on th e chart.
adults o f age 62 and older, more than 40% have pushed b. Are any points outside the control limits? Does it appear
back their retirement date? that the process is in control? Explain.
c. Comment on the difference between the tw o estimated 59. Acceptance sampling is an im portant quality control
probabilities. technique, where a batch of data is tested to determine if the
56. Presidential jo b approval is the most-watched statistic in proportion o f units having a particular attribute exceeds a
American politics. According to the June 2010 NBC/wall given percentage. Suppose that 10% of produced items are
known to be nonconforming. Every week a batch of items is Week Nonconforming Percentage
evaluated and the production machines are adjusted if the
1 5.5%
proportion of nonconform ing items exceeds 15%.
2 13.1%
a. What is the probability tha t the production machines
3 16.8%
w ill be adjusted if the batch consists o f 50 items?
4 13.6%
b. What is the probability tha t the production machines
5 19.8%
w ill be adjusted if the batch consists of 100 items?
6 2 .0 %
60. In the previous question, suppose the management decides
to use a chart for the analysis. As noted earlier, 10% of a. Suppose weekly batches consisted of 50 items.
produced items are known to be nonconforming. The C on stru ct a chart and determ ine if the machine need
firm analyzes a batch of production items for 6 weeks and adjustm ent in any o f the weeks.
computes the follow ing percentages o f nonconforming b. Suppose weekly batches consisted o f 100 items.
items. C o n stru ct a chart and determ ine if the machine needs
adjustm ent in any of the weeks.
CASE S T U D I E S
7.A1 Appendix
Derivation of the Properties of the Sample Mean
Let the expected value and the variance of the population X be denoted by E(X ) = μ and
Var(X ) = σ 2, respectively. The sample mean based on a random draw of n observations,
X 1, X 2, ., X n, from the population is computed as
We use the properties of the sum of random variables to derive
Since the sample mean is based on n independent draws from the population, the covari
ance terms drop out and the variance of the sample mean is thus derived as:
LEARNING OBJECTIVES
R
Estimation
After reading LO 8 .1 Discuss point estimators and their desirable properties.
this chapter
you should LO 8 .2 Explain an interval estimator.
be able to:
LO 8 .3 Calculate a confidence interval for the population mean when the population
standard deviation is known.
LO 8 .6 Calculate a confidence interval for the population mean when the population
standard deviation is not known.
LO 8 .8 Select a sample size to estimate the population mean and the population
proportion.
In earlier chapters we made a distinction between the population parameters, such as the p op u
lation mean and the population proportion, and the corresponding sample statistics. The sample
statistics are used to make statistical inferences regarding the unknown values o f the population
parameters. In general, tw o basic methodologies emerge from the inferential branch o f statistics:
estimation and hypothesis testing. In this chapter we focus on estimation. A point estimator is a
function o f the random sample that produces a single value as an estimate o f the unknown popula
tion parameter o f interest. A confidence interval, on the other hand, is a range o f values that is likely
to include the unknow n parameter. We first discuss p oint estimators and their desirable properties.
We then use point estimates to develop and interpret the confidence intervals o f the population
mean and the population proportion. Since obtaining a sample is one o f the first steps in making
statistical inferences, we also learn h ow an appropriate sample size is determined in order to achieve
a certain level o f precision in the estimates.
I N T R O D U C T O R Y CASE
FILE 97 117 93 79 97
87 78 83 94 96
102 98 82 96 113
113 111 90 101 99
112 89 92 96 98
In the above example, is a point estimate of the mean mpg for all ultra
green cars and is a point estimate of the proportion of all ultra-green cars thal
obtain over 100 mpg.
EXAMPLE 8 .1
A statistics section at a large university has 100 students. The scores of 10 randomly
selected final exams are:
66 72 40 85 75 90 92 60 82 38
U N B IA S E D N E S S
An estimator is unbiased if its expected value equals the unknown population
parameter being estimated.
Figure 8.1 shows the sampling distributions for two estimators U 1 and U2, which
are assumed to be normally distributed. Let θ (the Greek letter read as theta) be the true
parameter value of the population. Estimator U\ is unbiased because its expected value
E(U1) equals θ . Estimator U2 is biased because E ( U2) ≠ θ; the amount of bias is given by
the difference between E ( U 2) and θ .
Recall from Chapter 7 that and therefore, and are the unbi
ased estimators of μ and p , respectively. This property is independent of the sample size.
For instance, the expected value of the sample mean is equal to the population mean ir
respective of the sample size.
We often compare the performance of the unbiased estimators in terms of their relative
efficiency. An estimator is deemed efficient if its variability between samples is smaller
than that of other unbiased estimators. Recall that the variability is often measured by the
standard error of the estimator, where the standard error is the estimated standard devia
tion of the estimator. For an unbiased estimator to be efficient, its standard error must be
lower than that of other unbiased estimators. It is well documented that the estimators
and are not only unbiased, but also efficient.
EFFICIENCY
An unbiased estimator is efficient if its standard error is lower than that of other
unbiased estimators.
Figure 8.2 shows the sampling distributions for two unbiased estimators V 1 and V2 for the
true population parameter θ . Again, for illustration, V 1,and V2 follow the normal distribu
tion. While both V 1 and V2 are unbiased (E (V1) = E ( V 2) = θ), V 1 is more efficient because
it has less variability.
F IG U R E 8 . 2 T he d istrib u tio n s o f
efficient (V1) a n d less efficient (V2)
e stim a to rs
Another desirable property, which is often considered a minimum requirement for an esti
mator, is consistency. An estimator is consistent if it approaches the population parameter
of interest as the sample size increases. Consistency implies that we will get the inference
right if we take a large enough sample. The estimators and are not only unbiased, but
also consistent. For instance, the sample mean collapses to the population mean
as the sample size approaches infinity (n → ∞). An unbiased estimator is consistent if its
standard deviation, or its standard error, collapses to zero as the sample size increases.
CONSISTENCY
E X E R C I S E S 8. 1
Concepts Applications
1. A ra n d o m sa m p le p ro d u c e d th e follow ing data: 5. A ra n d o m sa m p le o f e ig h t d ru g sto re s show s th e follow ing
135 120 127 132 123 130 prices (in $) o f a p o p u lar pain reliever:
proportion o f those w ho believe that low crime is most a. Calculate the point estimate of the population
im portant. proportion o f those w ho would put the tax refund in
b. Calculate the point estimate o f the population the bank.
proportion o f those w ho believe tha t good jobs or b. Calculate the point estimate o f the population
affordable homes are most im portant. proportion of those w ho never get a refund.
C O N F ID E N C E IN T E R V A L
In order to construct a confidence interval for the population mean μ or the population pro
portion p, it is essential that the distributions of and follow, or approximately follow, a
normal distribution. Recall from Chapter 7 that follows a normal distribution when the
underlying population is normally distributed; this result holds irrespective of the sample
size n. If the underlying population is not normally distributed, then by the central limit
theorem, the sampling distribution of will be approximately normally distributed if the
sample size is sufficiently large, or when n ≥ 30. Similarly, the sampling distribution of
is approximately normally distributed if the sample size is sufficiently large, or when np
≥ 5 and n(1 – p ) ≥ 5.
A confidence interval is generally associated with a margin of error that accounts for
the variability of the estimator and the desired confidence level of the interval.
An analogy to a simple weather example is instructive. If you feel that the outside tem
perature is about 50 degrees, then perhaps you can, with a certain level of confidence, sug
gest that the actual temperature is between 40 and 60 degrees. In this example, 50 degrees
is analogous to a point estimate of the actual temperature and 10 degrees is the margin of
error that is added to and subtracted from this point estimate.
We know from the introductory case study that the point estimate of the population
mean mpg of all ultra-green cars is 96.5 mpg; that is, We can construct a con
fidence interval by using the point estimate as a base to which we add and subtract the
margin of error.
The symmetry of the normal distribution implies that the same margin of error will be
added to and subtracted from the point estimate. Since numerous samples of size n can be
drawn from the underlying population, the variability of the estimator is captured by its
standard deviation, or its standard error. The margin of error incorporates this important
information along with the desired confidence level of the interval.
LO 8.3
Constructing a Confidence Interval
Calculate a
confiden ce interval
for μ When σ Is Known
for t h e p o p u la t io n Consider a standard normal random variable Z. Using the symmetry of Z , we can compute
mean w hen the P(Z ≥ 1.96) = P(Z ≤ – 1.96) = 0.025; see Figure 8.4. Remember that z = 1.96 is easily
p o p u la t io n sta ndard determined from the z table given the probability of 0.025 in the upper-tail of the distribution.
d e vi a tio n is k n o w n . Therefore, we formulate the probability statement P(—1.96 ≤ Z ≤ 1.96) = 0.95.
This equation (see also the lower portion of Figure 8.4) implies that there is a 0.95 prob
ability that the sample mean will fall between and or
within the interval In other words, if samples of size n are drawn repeat
edly from a given population, 95% of the computed sample means, will fall within the
interval and the remaining 5% will fall outside the interval.
We do not know the population mean μ , and therefore cannot determine if a particular
falls within the interval or not. However, we do know that will fall within the interval
if and only if μ falls within the interval This will happen
95% of the time given how the interval is constructed. Therefore, we call the interval
a 95% confidence interval for the population mean, where is its
margin of error.
Confidence intervals are often misinterpreted; you need to exercise care in charac
terizing them. For instance, the above 95% confidence interval does not imply that the
probability that μ falls in the interval is 0.95. Remember that μ is a constant, although its
value is not known. It either falls in the interval (probability equals one) or does not fall in
the interval (probability equals zero). The randomness comes from not μ , since many
possible sample means can be derived from a population. Therefore, it is incorrect to say
that the probability that μ falls in the interval is 0.95. A 95% confidence
interval simply implies that if numerous samples of size n are drawn from a given popu
lation, then 95% of the intervals formed by the above formula will contain μ . Keep in
mind that we only use one sample to derive the estimates. Since there are many possible
samples, we will be right 95% of the time, thus giving us 95% confidence.
EXAMPLE 8.2
A sample of 25 cereal boxes of Granola Crunch, a generic brand of cereal, yields a
mean weight of 1.02 pounds of cereal per box. Construct a 95% confidence interval
of the mean weight of all cereal boxes. Assume that the weight is normally distrib
uted with a population standard deviation of 0.03 pound.
SOLUTI ON: Note that the normality condition of is satisfied since the underlying
population is normally distributed. A 95% confidence interval of the population
mean is computed as
With 95% confidence, we can report that the mean weight of all cereal boxes falls
between 1.008 and 1.032 pounds.
While it is common to report a 95% confidence interval, in theory we can construct an in
terval of any level of confidence ranging from 0 to 100%. Let’s now extend the analysis to
include intervals of any confidence level. Let the Greek letter α (alpha) denote the prob
ability of error, also known as the level of significance. This is the allowed probability that
the estimation procedure will generate an interval that does not contain μ . The confidence
coefficient (1 – α ) is interpreted similarly. The probability of error α and the confidence
level are related as
• α = 1 – confidence coefficient, and
• Confidence level = 100(1 — α ) % .
For example, a confidence coefficient of 0.95 implies that the probability of error α equals
1 - 0.95 = 0.05 and the confidence level equals 100(1 – 0.05)% = 95%. Similarly, for a
90% confidence interval, α = 1 – 0.90 = 0.10. The following statement generalizes the
construction of a confidence interval for μ when σ is known.
C O N F ID E N C E IN T E R V A L FOR μW H E N σ IS K N O W N
The notation zα /2 is the z value associated with the probability of α / 2 in the upper-tail
of the standard normal probability distribution, or P(Z > z α /2) = α / 2 . In other words,
if Z is a standard normal random variable and a is any probability, then z α /2 represents a
z value such that the area under the z curve to the right of zα/2 is α / 2 . Figure 8.5 depicts
the notation zα/2.
FIGURE 8.5 Graphical depiction o f the no tation zα /2
As discussed earlier, for a 95% confidence interval, α = 0.05 and α / 2 = 0.025. There
fore, Zα /2 = Z0.025 = 1.96. Similarly, we can derive the following:
• For a 90% confidence interval, α = 0.10, α / 2 = 0.05 and zα /2 = Z0.05 = 1.645.
• For a 99% confidence interval, α = 0.01, α / 2 = 0.005 and zα/2 = Z0.005 = 2.575.
I. For a given confidence level 100(1 — α )% and sample size n , the width of the
interval is wider, the greater the population standard deviation σ .
EXAMPLE 8.2b
Let the standard deviation of the population in Example 8.2 be 0.05 instead of 0.03.
Compute a 95% confidence interval based on the same sample information.
SOLU TION: We use the same formula as before, but we substitute 0.05 for the
standard deviation instead of 0.03:
II. For a given confidence level 100( 1 – α )% and population standard deviation σ ,
the width of the interval is wider, the smaller the sample size n.
EXAMPLE 8.2c
Instead of 25 observations, let the sample in Example 8.2 be based on 16 observa
tions. Compute a 95% confidence interval using the same sample mean of 1.02
pounds and the same population standard deviation of 0.03.
SOLU TION : Again we use the same formula as before, but this time we substitute
16 for n instead of 25:
EXAMPLE 8 .2d
Compute a 99%, instead of a 95%, confidence interval based on the information in
Example 8.2.
SOLUTION: Now we use the same formula and substitute the value 2.575 for zα /2
instead of 1.96:
The precision is directly linked with the width of the confidence interval—the wider the
interval, the lower is its precision. Continuing with the weather analogy, a temperature
estimate of 40 to 80 degrees is imprecise because the interval is too wide to be of value.
Examples 8.2b and 8.2c suggest that the estimate will be less precise if the variability of
the underlying population is high (σ is high) or a small segment of the population is sam
pled (n is small). Example 8.2d relates the width with the confidence level. For a given
sample information, the only way we can gain confidence is by making the interval wider.
If you are 95% confident that the outside temperature is between 40 and 60, then you can
increase your confidence level to 99% only by using a wider range, say between 35 and
65. This result also helps us understand the difference between precision (width of the
interval) and the confidence level. There is a trade-off between the amount of confidence
we have in an interval and its width.
EXAMPLE 8 .3
IQ tests are designed to yield results that are approximately normally distrib
uted. Researchers think that the population standard deviation is 15. A reporter is
interested in estimating the average IQ of employees in a large high-tech firm in
California. She gathers the IQ information on 22 employees of this firm and records
the sample mean IQ as 106.
a. Compute 90% and 99% confidence intervals of the average IQ in this firm.
b. Use these results to infer if the mean IQ in this firm is significantly different
from the national average of 100.
SOLUTION:
a. For a 90% confidence interval, zα /2 = Z o .o 5 = 1.645. Similarly, for a 99% confi
dence interval, za/2 = Z o .oo 5 = 2.575.
The 90% confidence interval is
The 99% confidence interval is
Note that the 99% interval is wider than the 90% interval.
b. The reporter wishes to determine if the mean IQ in this firm is significantly
different from the national average of 100. With 90% confidence, she can infer
that the average IQ of this firm’s employees differs from the national average,
since the value 100 falls outside the 90% confidence interval, [100.74, 111.26].
However, she cannot infer the same result with 99% confidence, since the
wider range of the interval, [97.77, 114.23], includes the value 100. We will
study the link between estimation and testing in more detail in the next chapter.
EXERCISES 8.2
Concepts 15. In order to e stim ate th e m ea n 30-year fixed m o r tg a g e
rate for a h o m e loan in th e United States, a random
9. Find z α /2 for each of the following confidence levels used in
s am p le o f 28 recent loans is taken. The average calculated
estimating the population mean.
from this sa m p le is 5.25%. It can be a ss u m e d that 30-year
a. 90%
fixed m o r tg a g e rates are normally distributed with a
b. 98% standard deviation of 0.50%. C om p ute a 90% and a 99%
c. 88% c o n fid e n c e interval for th e pop ulation m ean 30-year fixed
10. Find zα /2 for each of the following confidence levels used in m o r tg a g e rate.
estimating the population mean. 16. An article in the N a tio n a l Geographic News ("U.S. Racking
a. 89% Up Huge Sleep Debt,"February 2 4 , 2005) argues that
b. 92% Americans are increasingly skimping on their sleep. A
As discussed earlier, in order to derive a confidence interval of μ , it is essential that Discuss fea tu re s o f
the t distribution.
be normally distributed. A normally distributed is standardized as where
Z follows a standard normal distribution, that is, the z distribution. Another standard
ized statistic, which uses the estimator S in place of σ , is computed as The
random variable T follows the Student’s t distribution, more commonly known as the t
distribution.1
THE t D I S T R I B U T I O N
If repeated samples of size n are taken from a normal population with a finite vari
ance, then the statistic follows the t distribution with (n – 1) degrees of
freedom, df.
The t distribution is actually a family of distributions, which are similar to the z distribu
tion in that they are all bell-shaped and symmetric around zero. However, the t distribu
tion has slightly broader tails than does the z distribution. Each t distribution is identified
by the degrees of freedom, or simply df. The degrees of freedom determine the extent of
the broadness of the tails of the distribution; the fewer the degrees of freedom, the broader
the tails. Since the t distribution is defined by the degrees of freedom, it is common to
refer to it as the tdf distribution.
Specifically, the degrees of freedom refer to the number of independent pieces of in
formation that go into the calculation of a given statistic, and in this sense, can be “freely
chosen.” Consider the number of independent observations that enter into the calculation
of the sample mean. If it is known that and three of the observations have values
of x 1 = 16, x2 = 24, and x 3 = 18, then there is no choice but for the fourth observation to
have a value of 22. In other words, three degrees of freedom are involved in computing
if n = 4 ; in effect, one degree of freedom is lost.
'William S. Gossett (1876– 1937) published his research concerning the t distribution under the pen name “Student”
because his employer, the Guinness Brewery, did not allow employees to publish their research results.
• The tdf distribution has slightly broader tails than the z distribution.
• The tdf distribution consists of a family of distributions where the actual shape of
each one depends on the degrees of freedom df. As d f increases, the tdf distribution
becomes similar to the z distribution; it is identical to the z distribution when d f
approaches infinity.
From Figure 8.6 we note that the tails of the t2 and t5 distributions are broader than the
tails of the t50 distribution. For instance, for t2 and t5, the area exceeding a value of 3, or
P(Tdf > 3), is greater than that for t50. In addition, the t50 resembles the z distribution.
We use the notation tα,df to denote a value such that the area in the upper-tail equals α for
a given df. In other words, for a random variable Tdf, the notation ta,df represents a value
such that P(Tdf > ta,df) = α . Similarly, tα /2,df represents a value such that P(Tdf ≥ tα/ 2,df) =
α /2 . Figure 8.7 illustrates the notation.
Sometimes the exact probability cannot be determined from the t table. For example,
given d f = 10, the exact probability P(Tl0 ≥ 1.562) is not included in the table. However,
this probability is between 0.05 and 0.10 because the value 1.562 falls between 1.372 and
1.812. Similarly, P(T10 < 1.562) is between 0.90 and 0.95. We can use Excel and other
statistical packages to compute exact probabilities.
EXAMPLE 8 .4
Compute tα,df for α = 0.025 using 2, 5, and 50 degrees of freedom.
SOLUTION:
• For d f = 2, t0.025,2 = 4.303.
• For d f = 5, t0.025,10 = 2.571.
• For d f = 50, t0.025,60 = 2.009.
Note that the tdf values change with the degrees of freedom. Further, as d f increases,
the tdf distribution begins to resemble the z distribution. In fact with d f = ∞ ,
t0.025,∞ =o = 1.96, which is identical to the corresponding z value; recall that
P(Z ≥ 1.96) = 0.025.
As before, 100(1 – a)% is the confidence level and tα /2,df is the tdf value associated with
the probability α / 2 in the upper-tail of the distribution with d f = n – 1. In other words,
P(Tdf > tα/2,df) = α /2.
EXAMPLE 8.5
In the introductory case of this chapter, Jared Beane wants to estimate the mean
mpg of all ultra-green cars. Table 8.1 lists the mpg of a sample of 25 cars. Use this
information to construct a 90% confidence interval of the population mean. Assume
that mpg follows a normal distribution.
Thus, Jared concludes with 90% confidence that the average mpg of all ultra-green
cars is between 92.86 mpg and 100.18 mpg. Note that the manufacturer’s claim that
the ultra-green car will average 100 mpg cannot be rejected by the sample data,
since the value 100 falls within the 90% confidence interval.
D. Table 8.3 presents the relevant Excel output. As mentioned in Chapter 3, Excel
provides numerous numerical measures; however, we only include the point
estimate (Mean) and the margin of error (see the value next to Confidence Level
(90.0%)), or 96.52 ± 3.66, which is essentially the interval that we calculated
manually in Example 8.5.
TABLE 8.3 Confidence Interval for MPG Using Excel
Mean 96.52
EXERCISES 8.3
26. According to a recent survey, high school girls average c. How can Sarah reduce the margin of error reported in
100 text m essages daily (The Boston Globe, April 2 1 , 2010). the above interval estimate?
Assume that the survey was based on a random sample 30. A price-earnings ratio or P/E ratio is calculated as a firm's
of 36 high school girls.The sample standard deviation is share price compared to the income or profit earned by
c om puted as 10 text m essages daily. the firm per share. Generally, a high P/E ratio suggests that
a. Calculate the margin of error with 99% confidence. investors are expecting higher earnings growth in the future
b. What is the 99% confidence interval of the population compared to companies with a lower P/E ratio. The following
mean texts that all high school girls send daily? table sh ows the P/E ratios for a sample of firms in the
footwear industry:
27. The Chartered Financial Analyst (CFA®) designation is
fast becoming a requirement for serious investment Firm P/E Ratio
professionals. Although it requires a successful completion Brown Shoe Co., Inc. 20.54
of three levels of grueling exams, it also entails promising
Collective Brands, Inc. 9.33
careers with lucrative salaries. A student of finance is curious
CROCS Inc. 22.63
about the average salary of a CFA® charterholder. He takes a
random sample of 36 recent charterholders and computes DSW Inc. 14.42
Source: http://finance.yahoo.com .
C O N F I D E N C E I N T E R V A L FOR p
A 100(1 − α )% confidence interval of the population proportion p is computed as
EXAMPLE 8 .6
In the introductory case of this chapter, Jared Beane wants to estimate the proportion
of all ultra-green cars that obtain over 100 mpg. Use the information in Table 8.1 to
construct a 90% and a 99% confidence interval of the population proportion.
SOLUTION: As shown in Table 8.1, seven of the 25 cars obtain over 100 mpg; thus the
point estimate of the population proportion is Note that the
normality condition is satisfied since np ≥ 5 and n ( 1 − p) ≥ 5, where p is evaluated
With a 90% confidence level, α / 2 = 0.10/2 = 0.05; thus we find
Za/2 = Z0.05 = 1.645. Substituting the appropriate values into
With 90% confidence, Jared reports that the percentage of cars that obtain over
100 mpg is between 13.2% and 42.8%.
If Jared had wanted a 99% confidence level, we would use α / 2 = 0.01/2 = 0.005
and Za/ 2 = Zo.oo5 =: 2.575 to obtain
At a higher confidence level of 99%, the interval for the percentage of cars that
obtain over 100 mpg becomes 4.9% to 51.1%. Given the current sample size of 25
cars, Jared can gain confidence (from 90% to 99%) at the expense of precision, as
the corresponding margin of error increases from 0.148 to 0.231.
EXERCISES 8.4
Concepts 36. Assume in a sample size o f n = 50.
a. Construct a 95% confidence interval o f the population
34. A random sam ple o f 100 observations results in 40 successes.
proportion.
a. What is the point estimate of the population proportion
b. What happens to the margin o f error if the above sample
of successes?
proportion is based on n = 200 instead of n = 50?
b. Construct a 90% confidence interval of the population
proportion. 37. A sam ple o f 80 results in 30 successes.
c. Construct a 99% confidence interval o f the population a. Calculate the point estimate o f the population
c. What happens to the margin of error as you move from 44. In an NBC N ew s/W all Street Journal poll of 1,000 American
an 88% confidence interval to a 98% confidence interval? adults conducted on August 5–9 , 2 0 1 0 , 44% of respondents
approved of the job that Barack Obama was doing in
Applications handling th e economy.
39. A survey of 1,026 people asked: "What would you do with an a. Compute a 90% confidence interval of the proportion of
unexpected tax refund?" Forty-seven percent responded that Americans w h o approved o f Barack Obama's handling of
they would pay off debts (Vanity Fair, June 2010). the economy.
a. At 95% confidence, what is the margin of error? b. What is the resulting margin of error?
b. Construct a 95% confidence interval of the population c. Compute the margin of error associated with a 99%
proportion of people w h o would pay off debts with an confidence level.
unexpected tax refund. 45. Obesity is generally defined as 30 or more pounds over a
40. A recent poll of 1,079 adults finds that 51% of Americans healthy weight. A recent study of obesity reports 27.5% of a
support Arizona's stringent new immigration enforcement random sample of 400 adults in the United States to be obese.
law, even though it may lead to racial profiling (N ew York a. Use this sample information to compute a 90% confidence
Times/CBS News, April 28-M ay 2 , 2010). Use the sample interval of the adult obesity rate in the United States.
information to com pute a 95% confidence interval of the b. Is it reasonable to conclude with 90% confidence that
population parameter of interest. the adult obesity rate in the United States is not 30%?
41. An economist reports that 560 out of a sample of 1,200 46. An accounting professor is notorious for being stingy in
middle-income American households actively participate in giving out g oo d letter grades. In a large section of 140
the stock market. students in the fall semester, she gave out only 5% As,
a. Construct a 90% confidence interval of the proportion 23% Bs, 42% Cs, and 30% Ds and Fs. Assuming that this was
of middle-income Americans w h o actively participate in a representative class, com pute a 95% confidence interval of
the stock market. the probability of getting at least a B from this professor.
b. Can w e conclude that the proportion of middle-income 47. One in five 18-year-old Americans has not graduated from
Americans that actively participate in the stock market is high school (The Wall Street Journal, April 1 9 , 2007). A mayor
not 50%? of a northeastern city com ments that its residents do not
42. In a CNNMoney.com poll condu cted on July 1 3 , 2010, a have the sam e graduation rate as the rest of the country. An
sample of 5,324 Americans were asked about what matters analyst from the Department of Education decides to test the
m ost to them in a place to live. Thirty-seven percent of the mayor's claim. In particular, she draws a random sample of
respondents felt job opportunities matter most. 80 18-year-olds in the city and finds that 20 of them have not
a. Construct a 90% confidence interval of the proportion of graduated from high school.
Americans w h o feel that go od job opportunities matter a. Compute the point estimate of the proportion of
most in a place to live. 18-year-olds w h o have graduated from high school in
b. Construct a 99% confidence interval of the proportion of this city.
Americans w h o feel that g ood job opportunities matter b. Use this point estimate to derive a 95% confidence
most in a place to live. interval of the population proportion.
c. Which of the above tw o intervals has a higher margin of c. Can the mayor's com m ent be justified at 95%
error? Explain why. confidence?
Selecting n to Estimate µ
Consider a confidence interval for µ with a known population standard deviation σ . In
addition, let D denote the desired margin of error. In other words, you do not want the
sample mean to deviate from the population mean by more than D , for a given level
we rearrange this equation to derive the formula for
of confidence. Since
the required sample size as The sample size can be computed if we specify
the population standard deviation σ , the value of zα/2 based on the confidence level
100(1 − α )%, and the desired margin of error D.
This formula is based on a knowledge of σ . However, in most cases σ is not known
and therefore, has to be estimated. In such cases, we replace σ with its reasonable
estimate
THE R EQ U I R E D S A M P L E SIZE W H E N
ES T IMA T IN G THE P O P U L A T I O N M E A N
For a desired margin of error D, the minimum sample size n required to estimate a
100(1 − α )% confidence interval of the population mean µ is
EXAMPLE 8.7
Let us revisit Example 8.5, where Jared Beane wants to construct a 90% confi
dence interval of the mean mpg of all ultra-green cars. Suppose Jared would like
to constrain the margin of error to within 2 mpg. Further, Jared knows that the
lowest mpg in the population is 76 mpg, whereas the highest is 118 mpg. How
large a sample does Jared need to compute the 90% confidence interval of the
population mean?
SOLUTION: For a 90% confidence level, Jared computes zα /2 = Z0.05 = 1.645. He
estimates the population standard deviation as
10.50. Given D = 2, the required sample size is
Selecting n to Estimate p
The margin of error D of the confidence interval of the population proportion p is D =
represents the sample proportion. By rearranging, we derive the
formula for the required sample size as Note that this formula is not
feasible because it uses which cannot be computed unless a sample of size n has already
been selected. We replace with a reasonable estimate of the population proportion p.
THE R E Q U I R E D S A M P L E SIZE W H E N E S T I MA T I N G
THE P O P U L A T I O N P R O P O R T I O N
For a desired margin of error D, the minimum sample size n required to estimate a
100(1 − α )% confidence interval of the population proportion p is
Sometimes we use the sample proportion from a preselected sample as in the plan
ning stage. Once the optimal sample size is determined, the final sample is selected for
estimating the population proportion. Another choice for is to use an estimate of the
population proportion from prior studies. If no other reasonable estimate of the popula
tion proportion is available, we can use as a conservative estimate to derive the
optimal sample size; note that the required sample is the largest when
EXAMPLE 8 .8
Recall Example 8.6, where Jared Beane wants to construct a 90% confidence inter
val of the proportion of all ultra-green cars that obtain over 100 mpg. Jared does not
want the margin of error to be more than 0.10. How large a sample does Jared need
for his analysis of the population proportion?
which is rounded up to 68. Therefore, Jared needs to find another random sample of
at least 68 ultra-green cars to provide a more precise interval estimate of the propor
tion of all ultra-green cars that obtain over 100 mpg.
SYNOPSIS OF I N T R O D U C T O R Y CASE
Jared B e a n e , a j o u r n a lis t a t a r e p u t a b l e a u t o m o b i l e m a g a z i n e , p r e
p a r e s t o w r i t e a n article o n t h e n e w u l t r a - g r e e n car t h a t b o a s t s an
a v e r a g e o f 1 0 0 m p g . B a s e d o n a s a m p l e o f 2 5 cars, Jared r e p o r ts
w i t h 9 0 % c o n f i d e n c e t h a t t h e a v e r a g e m p g o f all u l t r a - g r e e n cars
is b e t w e e n 9 2 . 8 6 m p g a n d 1 0 0 . 1 8 m p g . Jared a l s o c o n s t r u c t s a
9 0% c o n f i d e n c e interval for t h e p r o p o r t i o n o f cars t h a t o b t a i n
m o r e t h a n 1 0 0 m p g a n d o b t a i n s a n in te rval e s t i m a t e o f 0 . 1 3 2 t o
0 .4 2 8 . Jared w i s h e s t o i n c r e a s e t h e p r e c i s i o n o f his c o n f i d e n c e in
te r v a l s b y r e d u c i n g t h e m a r g in o f error. If his d e s i r e d m a r g in o f
error is 2 m p g for t h e p o p u l a t i o n m e a n , h e m u s t u s e a s a m p l e
o f a t l e a s t 7 5 cars for t h e a n aly sis. Jar ed a l s o w a n t s t o r e d u c e t h e
m a r g in o f error t o 0 . 1 0 for t h e p r o p o r t i o n o f cars t h a t o b t a i n m o r e
t h a n 1 0 0 m p g . U s in g a c o n s e r v a t i v e e s t i m a t e , h e c a l c u l a t e s t h a t a
s a m p l e o f a t le a s t 6 8 cars is n e e d e d t o a c h i e v e th is g o a l . T hus, in o r d e r t o g a i n p r e c i s i o n in
t h e interval e s t i m a t e o f b o t h t h e m e a n a n d t h e p r o p o r t i o n w i t h 9 0 % c o n f i d e n c e , Jared's
s a m p l e m u s t c o n t a i n a t l e a s t 7 5 cars.
EXERCISES 8.5
Concepts What happens to n if you decide to estimate p with 90%
confidence?
48. What is the minimum sample size n required to estimate
µ with 90% confidence if the desired margin of error is 54. You wish to com pute a 95% confidence interval of the
D = 1.2? The population standard deviation is estimated population proportion. How large a sample should you
What happens to n if the desired margin of error draw to ensure that the sample proportion does not deviate
decreases to D = 0.7? from the population proportion by more than 0.06? No prior
estimate of the population proportion is available.
49. The lowest and highest observations in a population are
20 and 80, respectively. What is the minimum sample size n
required to estimate µ with 80% confidence if the desired Applications
margin of error is D = 2.6? What happens to n if you decide
55. An analyst from an energy research institute in California
to estimate µ with 95% confidence?
wish es to precisely estimate a 99% confidence interval of the
50. Find the required sample size for estimating the population average price of unleaded gasoline in the state. In particular,
mean in order to be 95% confident that the sample mean she d oes not want the sample mean to deviate from the
is within 10 units of the population mean. Assume that the population mean by more than $0.06. What is the minimum
population standard deviation is 40. number of gas stations that she should include in her sample
51. You need to com pute a 99% confidence interval of the if she uses the standard deviation estimate of $0.32, as
population mean. How large a sample should you draw to reported in the popular press?
ensure that the sample mean do es not deviate from the 56. An analyst would like to construct 95% confidence intervals
population mean by more than 1.2? (Use 6.0 as an estimate of the mean returns of tw o mutual funds. Fund A is a high-
of the population standard deviation from prior studies.) risk fund with a known population standard deviation of
52. What is the minimum sample size n required to estimate p 20.6%, whereas the Fund B is a lower-risk fund with a known
with 95% confidence if the desired margin of error D = 0.08? population standard deviation of 12.8%.
The population proportion is estimated as a. What is the minimum sample size required by the
prior studies. What hap pens to n if th e desired margin of analyst if she wants to restrict the margin of error to 4%
error increases to D = 0.12? for Fund A?
53. In the planning stage, a sample proportion is estimated b. What is the minimum sample size required by the analyst if
as Use this information to compute she wants to restrict the margin of error to 4% for Fund B?
the minimum sample size n required to estimate p with c. Why do the above results differ if they use the sam e
99% confidence if the desired margin of error D = 0.12. margin o f error?
57. The manager of a pizza chain in Albuquerque, New Mexico, a. Identify the relevant parameter o f interest for these
wants to determine the average size of their advertised qualitative data and compute its point estimate as well
16-inch pizzas. She takes a random sample o f 25 pizzas and as the margin of error with 90% confidence.
records their mean and standard deviation as 16.10 and b. You decide to redo the analysis with the margin of error
1.8 inches, respectively. She subsequently com putes a 95% reduced to 2%. How large a sample do you need to
confidence interval of the mean size of all pizzas as [15.36, draw? State your assumptions in computing the required
16.84], However, she finds this interval to be too broad sample size.
to implement quality control and decides to re-estimate
60. Subprime lending was big business in the United States in
the mean based on a bigger sample. Using the standard
the mid-2000s, w hen lenders provided m ortgages to people
deviation estimate of 1.8 from her earlier analysis, how large
with poor credit. However, su bsequent increases in interest
a sample must she take if she wants the margin of error to be
rates coupled with a drop in hom e values necessitated many
under 0.5 inch?
borrowers to default. Suppose a recent report finds that two
58. Mortgage lenders often use FICO® scores to check the credit in five subprime mortgages are likely to default nationally.
worthiness of consumers applying for real estate loans. A research econom ist is interested in estimating default
In general, FICO scores range betw een 300 and 850 with rates in Illinois with 95% confidence. How large a sample is
higher scores representing a better credit profile. A lender in needed to restrict the margin of error to within 0.06, using
a Midwestern town would like to estimate the mean credit the reported national default rate?
score of its residents. What is the required number of sample
61. A student of business is interested in estimating a 99%
FICO scores needed if the lender does not want the margin of
confidence interval of the proportion o f students who
error to exceed 20, with 95% confidence?
bring laptops to campus. He wishes a precise estimate and
59. A survey by the AARP (Money, June 2007) reported that is willing to draw a large sample that will keep the sample
approximately 70% of people in the 50 to 64 age bracket proportion within five percentage points of the population
have tried so m e type of alternative therapy (for instance, proportion. What is the minimum sample size required by
acupuncture or the use of nutrition supplements). Assume this student, given that no prior estimate of the population
this survey was based on a sample of 400 people. proportion is available?
Soufce: http://finance.yahoo.com.
Sample Grim news continues to distress the housing sector. On August 24, 2010, Reuters reported
that the sales of previously owned U.S. homes took a record plunge in July to the slow
Report— est pace in 15 years. Combine this fact with the continued fallout from the subprime
mortgage debacle, a sluggish economy, and high unemployment, and the housing sector
Weekly Stock appears quite unstable. Have these unfavorable events managed to trickle down and harm
the financial performance of Home Depot and Lowe’s, the two largest home improvement
Performance: retailers in the U.S.?
One way to analyze their financial stability is to observe their stock performance dur
Home Depot ing this period. In order to make valid statements concerning the reward of holding these
stocks, weekly return data for each firm were gathered from January through August of
vs. Lowe's 2010. Table 8.A summarizes the important descriptive statistics.
TABLE 8 . A Descriptive Statistics for Weekly Returns o f Hom e D e p o t and Lowe's (n = 34)
Over the past 34 weeks, Home Depot posted both the higher average return and me
dian return of 0.00% and 0.76%, respectively. Lowe’s return over the same period was
negative, whether the central tendency was measured by its mean ( − 0.33%) or its median
( − 0.49%). In terms of dispersion, Lowe’s return data had the higher standard deviation
(3.83% > 3.59%). In terms of descriptive measures, the investment in Home Depot’s
stock not only provided higher returns, but was also less risky than the investment in
Lowe’s stock.
Table 8.A also shows the margins of error for 95% confidence intervals for the mean
returns. With 95% confidence, the mean return for Home Depot fell in the range [− 1.25%,
1.25%], while that for Lowe’s fell in the range [− 1.67%, 1.01%]. Given that these two
intervals overlap, one cannot conclude that Home Depot delivered the higher reward over
this period— a conclusion one may have arrived at had only the point estimates been
evaluated. It is not possible to recommend one stock over the other for inclusion in a cli
ent’s portfolio based solely on the mean return performance. Other factors, such as the
standard deviation of the returns and the correlation between the stock and the existing
portfolio, must be analyzed before this decision can be made.
Conceptual Review
LO 8.1 Discuss p o in t estim ators an d th e ir desirable properties.
The sample mean is the point estimator of the population mean µ and the sample
proportion is the point estimator of the population proportion p. Sample values of the
point estimators represent the point estimates of the population parameter of interest;
are the point estimates of µ and p , respectively.
Some of the desirable properties of point estimators are unbiasedness, efficiency, and
consistency. Unbiasedness indicates that no systematic bias appears against the un
known population parameter of interest; in other words, the expected value of the
estimator equals the true value of the population parameter. An unbiased estimator is
deemed efficient if its variability is less than that of other unbiased estimators. Con
sistency implies that the estimator approaches the population parameter as the sample
size gets very large. It is well documented that are the unbiased, efficient, and
consistent estimators of µ and p, respectively.
LO 8 .8 Select a sample size to estimate the population mean and the population proportion.
For a desired margin of error D, the minimum n required to estimate p. with
100(1 — a)% confidence is is a reasonable estimate of σ in
the planning stage. If cr is known, we replace with cr. Other choices for include
an estimate from a preselected sample, prior studies, or For a desired
margin of error D, the minimum n required to estimate p with 100(1 − α )% confidence
is a reasonable estimate of p in the planning stage.
Choices for include an estimate from a preselected sample or prior studies; a con
servative estimate of is used when no other reasonable estimate is available.
64. A hair salon in Cambridge, Massachusetts, reports that 67. A machine that is programmed to package 1.20 pounds of
on seven randomly se lected weekdays, the number of cereal is being tested for its accuracy. In a sample of 36 cereal
customers w h o visited the salon were 40, 3 0 , 2 8 , 2 2 , 3 6 , 16, boxes, the sample mean filling w eight is calculated as 1.22
and 50. It can be assumed that weekday customer visits pounds. The population standard deviation is known to be
follow a normal distribution. 0.06 pound.
a. Construct a 90% confidence interval of the average a. Identify the relevant parameter of interest for these
number of customers w h o visit the salon on weekdays. quantitative data and com pute its point estimate as well
b. Construct a 99% confidence interval of the average as the margin of error with 95% confidence.
number of customers w h o visit the salon on weekdays. b. Can w e conclude that the packaging machine is
c. What happens to the width of the interval as the operating properly?
confidence level increases? c. How large a sample must w e take if w e wish to contain
the margin of error below 0.01 pound with 95%
65. Recently, six single-family ho m es in San Luis Obispo County
in California sold at the following prices (in $1,000s): $549, confidence?
$449, $705, $529, $639, $609. 68. The SAT is the most widely used test in the undergraduate
a. Construct a 95% confidence interval for the mean sale admissions process. Scores on the math portion o f the SAT
price in San Luis Obispo County. are believed to be normally distributed and range from
200 to 800. A researcher from the admissions department 72. According to a survey of 1,235 businesses by IDC, a market-
at the University of New Hampshire is interested in research concern in Framingham, Massachusetts, 12.1% of
estimating the mean math SAT scores of the incoming class sole proprietors are engaging in e-com m erce ( The Wall Street
with 90% confidence. How large a sample should she take to Journal, July 2 6 , 2007).
ensure that the margin of error is below 15? a. What is the margin of error in estimating with 95%
69. A recent study by Allstate Insurance Co. finds that 82% of confidence the proportion of sole proprietors that
teenagers have used cell phones while driving ( The Wall e n g a g e in e-commerce?
Street Journal, May 5 , 2010). Suppose this study was based on b. Construct a 95% confidence interval of the population
a random sample of 50 teen drivers. proportion.
a. Construct a 99% confidence interval for the proportion 73. A Monster.com poll of 3,057 individuals asked: "What's
of all teenagers that have used cell phones while the longest vacation you plan to take this summer?" The
driving. following relative frequency distribution summarizes the
b. What is the margin o f error with 99% confidence? results.
70. The following table shows the annual returns (in percent) Response Relative Frequency
for the Vanguard Energy Fund.
A few days 21%
Year Return A few long weekends 18%
2005 44.60 One week 36%
2006 19.68 Two weeks 22 %
2007 37.00
Source: The Boston Globe, June 12,2007.
2008 − 42.87
a. Construct a 95% confidence interval for the proportion
2009 38.36
of people w h o plan to take a one-w eek vacation this
Source: h ttp ://fin a n c e .y a h o o .c o m . summer.
a. Calculate the point estimate for µ . b. Construct a 99% confidence interval for the proportion
b. Construct a 95% confidence interval for µ . of people w h o plan to take a one-w eek vacation this
summer.
c. What assumption did you make w hen constructing
the interval? c. Which of the two confidence intervals is wider?
71. While the housing market is in the tank and is not 74. Linda Barnes has learned from prior studies that one out
likely to em erg e anytime soon, real estate investment in o f five applicants gets admitted to top MBA programs in
college towns continues to promise g o o d returns ( The Wall the country. She wishes to construct her own 90%
Street Journal, S eptember 2 4 , 2010). Marcela Treisman works confidence interval of the acceptance rate in top MBA
for an investment firm in Michigan. Her assignm ent is to programs. How large a sample should she take if she
analyze the rental market in Ann Arbor, which is hom e to d o es not want the acceptance rate of the sample to deviate
the University of Michigan. She gathers data on monthly from that of the population by more than five percentage
rents for 2011 along with the square foo tag e o f 40 homes. points? State your assumptions in computing the required
A portion of the data is shown below; the com plete data, sample size.
labeled Ann Arbor Rental, can be found on the text website. 75. According to a recent report by the PEW Research
Center, 85% of adults under 30 feel optimistic about
Monthly Rent Square Footage
the economy, but the optimism is shared by only 45%
645 500 of those w h o are over 50 (Newsweek, September 13,
675 648 2010). A research analyst would like to construct 95%
confidence intervals for the proportion patterns in
2400 2700 various regions of the country. She uses the reported
rates by the PEW Research Center to determine the
Source: w w w .z illo w .c o m .
sample size that would restrict the margin of error to
a. Use Excel to calculate the mean and standard deviation within 0.05.
for monthly rent. a. How large a sample is required to estimate the
b. Use Excel to calculate the mean and standard deviation proportion o f adults under 30 w h o feel optimistic about
for square footage. the economy?
c. Use Excel to construct 90% confidence intervals for the b. How large a sample is required to estimate the
mean rent and the mean square footage of all hom es in proportion of adults over 50 w h o feel optimistic about
Ann Arbor, Michigan. the economy?
CASE STUDIES
Data fo r Case Study 8.1 Hourly W ages of Texas Workers by Education Level (in $)
Bachelor's Degree or Higher High School Diploma No High School Diploma
$22.50 $ 12.68 $ 11.21
Source: http://finance.yahoo.com.
In a report, use the above information to:
1. Use descriptive statistics to compare the returns of the mutual funds.
2. Assess reward by constructing and interpreting 95% confidence intervals for the
population mean return. What assumption did you make for the interval estimates?
Question: Which are more important to you in your vote for Congress this
November: domestic issues such as the economy, health care, and immigration;
or international issues such as Afghanistan, Iran, and terrorism?
In a report, construct 95% confidence intervals of the relevant population proportions to:
1. Compare the approval rates of President Obama’s handling of the Gulf Coast oil
spill and President George W. Bush’s handling of the Hurricane Katrina crisis.
2. Compare the importance of domestic issues in August 2010 and in September 2006.
Hypothesis Testing
C H A P T E R
LEARNING OBJECTIVES
After reading LO 9 .1 Define the null hypothesis and the alternative hypothesis.
this chapter
you should LO 9 .2 Distinguish between Type I and Type II errors.
be able to:
LO 9 .3 Explain the steps of a hypothesis test using the p-value approach.
LO 9 .4 Explain the steps of a hypothesis test using the critical value approach.
In Chapter 8, we learned how to estimate an unknow n population parameter o f interest using point
estimates and confidence intervals. In this chapter, we will focus on the second major area o f sta
tistical inference, called hypothesis testing. We use a hypothesis test to challenge the status quo, or
some belief about an underlying population parameter, based on sample data. For instance, we may
wish to test w hether the average age o f MBA students in the U.S. is less than 30 years. Since we do
not have access to the ages o f all MBA students in the country, we have to perform statistical infer
ence on the basis o f limited sample information. Suppose that a sample o f 40 MBA students reveals
a mean age o f 29 years. Although the sample mean is less than 30, it does not guarantee that the
mean age o f all MBA students in the population is less than 30. We may be able to justify the lower
sample mean by pure chance. In this chapter, we will discuss how to determine w hether the conclu
sion from the sample can be deemed real (that is, the mean age is less than 30) or due to chance
(that is, the mean age is not less than 30).
I N T R O D U C T O R Y C A S E
TABLE 9 . 1 Average Hours Studied per Week for a Sample o f 35 College Students
25 17 8 14 17 7 11
19 16 9 15 12 17 19
26 14 22 17 14 35 24
11 21 6 20 27 17 6
29 10 10 4 25 13 16
S u s a n w a n t s t o u s e t h e s a m p l e i n f o r m a t io n to:
1. D e t e r m i n e if t h e m e a n s t u d y t i m e o f s t u d e n t s a t h e r u n iv e r sit y is b e l o w t h e 1961 n a t i o n a l a v e r a g e o f 2 4 h o u r s
per w e e k .
W e p r o v i d e a s y n o p s i s o f th is c a s e a t t h e e n d o f S e c t i o n 9.3.
9.1 In troduc tion to Hypothesis Testing
LO9 .1 We use hypothesis testing to resolve conflicts between two competing opinions (hypoth
Define th e null eses) on a particular population parameter of interest. We refer to one hypothesis as the
hypothesis and null hypothesis, denoted H0, and the other as the alternative hypothesis, denoted HA.
th e alternative We can think of a null hypothesis as corresponding to a presumed default state of nature
hypothesis. or status quo. An alternative hypothesis, on the other hand, contradicts the default state
or status quo. In other words, only one of the two hypotheses is true and the hypotheses
cover all possible values of the population parameter.
In statistics we use sample information to make inferences regarding the unknown popu
lation parameters of interest. In this chapter our goal is to determine if the null hypothesis
can be rejected in favor of the alternative hypothesis. An analogy can be drawn with
applications in the medical and legal fields, where we can define the null hypothesis as
“an individual is free of disease” or “an accused is innocent.” In both cases the verdict is
based on limited evidence, which in statistics translates into making a decision based on
limited sample information.
On the basis of sample information, we either “reject the null hypothesis” or “do
not reject the null hypothesis.”
Consider the example just referenced where the null is defined as “an individual is free of
disease.” Suppose a particular medical procedure does not detect the disease. On the basis
of this limited information, we can only conclude that we are unable to detect the disease
(do not reject the null hypothesis). It does not necessarily prove that the person does not
have the disease (accept the null hypothesis). Similarly, in the court example where the
null hypothesis is defined as “an accused is innocent,” we can conclude that the person is
guilty (reject the null hypothesis) or that there is not enough evidence to convict (do not
reject the null hypothesis).
A hypothesis test can be one-tailed or two-tailed. A two-tailed test is defined when the
null hypothesis states a specific value for the population parameter of interest. For exam
ple, H0: µ = µ0 versus HA: µ ≠ µ0 and H0: p = p0 versus HA: p ≠ p 0 are two-tailed tests,
where µ 0 and p 0 represent hypothesized values of the population mean and the population
proportion, respectively. If the null hypothesis is rejected, it suggests that the true param
eter does not equal the hypothesized value.
A one-tailed test, on the other hand, involves a null hypothesis that can only be
rejected on one side of the hypothesized value. For example, consider H0: µ ≤ µ0 versus
HA: µ > µ 0. Here we can reject the null hypothesis only when there is substantial evi
dence that the population mean is greater than µ 0. It is also referred to as a right-tailed
test since rejection of the null hypothesis occurs on the right side of the hypothesized
mean. Another example is a left-tailed test, H0: µ ≥ µ 0 versus HA: µ < µ 0, where the null
hypothesis can only be rejected on the left side of the hypothesized mean. One-tailed tests
for the population proportion are defined similarly.
EXAMPLE 9 .2
An advertisement for a popular weight-loss clinic suggests that participants in its
new diet program lose, on average, more than 10 pounds. A consumer activist wants
to determine if the advertisement’s claim is valid. Specify the null and the alterna
tive hypotheses to validate the advertisement’s claim.
SOLUTION: The advertisement’s claim concerns average weight loss, thus the
parameter of interest is again the population mean. This is an example of a one
tailed test because we want to determine if the mean weight loss is more than 10
pounds (µ > 10). We specify the competing hypotheses as
H0: µ ≤ 10 pounds
HA: µ > 10 pounds
The underlying claim that the mean weight loss is more than 10 pounds is true if
our decision is to reject the null hypothesis. Conversely, if we do not reject the null
hypothesis, we infer that the claim is not supported by the sample data.
EXAMPLE 9.3
A television research analyst wishes to test a claim that more than 50% of the house
holds will tune in for a TV episode. Specify the null and the alternative hypotheses
to test the claim.
T Y P E I A N D T Y P E II E R R O R S
A Type I error is committed when we reject the null hypothesis when the null
hypothesis is true. A Type II error is made when we do not reject the null hypoth
esis when the null hypothesis is false. The probability of a Type I error is denoted
by α and the probability of a Type II error is denoted by ß . For a given sample
size n, a decrease in α will increase ß and vice versa. Both α and ß decrease as
n increases.
EXAMPLE 9.5
Consider the following hypotheses that relate to the medical example mentioned
earlier.
H0: A person is free of disease
HA: A person has disease
Suppose a person takes a medical test that attempts to detect the disease. Discuss the
consequences of a Type I error and a Type II error.
SOLUTIO N: A Type I error occurs when the medical test indicates that the person
has the disease (reject H0), but in reality, the person is free of the disease. We often
refer to this type of result as a false positive. If the medical test shows that the person
is free of the disease (do not reject H0) when the person actually has the disease,
then a Type II error occurs. We often call this type of result a false negative.
EXAMPLE 9 .6
Consider the following competing hypotheses that relate to the court of law.
H0: An accused person is innocent
HA: An accused person is guilty
Suppose the accused person is judged by a jury of her peers. Discuss the conse
quences of a Type I error and a Type II error.
SOLUTION: A Type I error is a verdict that finds that the accused is guilty (reject
H0) when she is actually innocent. A Type II error is due to a verdict that concludes
the accused is innocent (do not reject H0) when in reality, she is guilty. For given
evidence, we cannot reduce either of these two errors without increasing the other.
Moreover, as mentioned earlier, it is not always easy to determine which of the two
errors is more costly to society.
Table 9.2 summarizes the circumstances surrounding Type I and Type II errors. Two cor
rect decisions are possible: not rejecting the null hypothesis when the null hypothesis is
true, and rejecting the null hypothesis when the null hypothesis is false. Conversely, two
incorrect decisions (errors) are also possible: rejecting the null hypothesis when the null
hypothesis is true (Type I error), and not rejecting the null hypothesis when the null hy
pothesis is false (Type II error).
As we will see in later sections, a hypothesis test is based on the chosen level of the Type I
error α . We want the test to have the ability to reject the null hypothesis when the null hy
pothesis is false. The power of the test is defined as the probability of rejecting the null hy
pothesis when the null hypothesis is false. Since ß is the probability of a Type II error, the
power of the test is defined as 1 – ß . In more advanced statistics, we learn how to compare
various types of tests on the basis of their size and power, where the size denotes the actual
proportion of rejections when the null is true and the power denotes the actual proportion
of rejections when the null is false. These analyses are based on computer simulations.
EXERCISES 9. 1
1. Explain why the following hypotheses are not constructed c. Test if Americans ge t an average of less than seven hours
correctly. of sleep.
a. H0: µ ≤ 10;0≥
1µ 6. Define the consequences of Type I and Type II errors for each
b. H0: µ 500;
≠ HA: µ = 500 of the tests considered in th e preceding question.
c. H0: p ≤ 0.40; HA: p > 0.42 7. Construct the null and alternative hypotheses for the
following claims:
2. Which o f the following statements are valid null and a. "I am going to ge t the majority of the votes to win this
alternative hypotheses? If they are invalid hypotheses, election."
explain why. b. "I susp ect that your 10-inch pizzas are, on average, less
than 10 inches in size."
b. H0: µ = 120; HA: µ ≠ 120 c. "I will have to fine the company since its tablets do
c. H0: p ≤ 0.24; HA: p > 0.24 not contain an average of 250 mg of ibuprofen as
advertised."
d. H0: µ
< 252; HA. µ > 252
8. Discuss the consequences of Type I and Type II errors for each
3. Explain why the following statements are not correct.
of the claims considered in the preceding question.
a. "With my methodological approach, I can reduce the
Type I error with the given sample information without 9. A polygraph (lie detector) is an instrument used to
changing the Type II error." determine if the individual is telling the truth. These tests are
considered to be 95% reliable. In other words, if an individual
b. "I have already decided how much o f the Type I error
lies, there is a 0.95 probability that the test will detect a lie.
I am going to allow. A bigger sample will not change
Let there also be a 0.005 probability that the test erroneously
either the Type I or Type II errors."
detects a lie even when the individual is actually telling the
c. "I can reduce the Type II error by making it difficult to
truth. Consider the null hypothesis, "the individual is telling
reject the null hypothesis."
the truth," to answer the following questions.
d. "By making it easy to reject the null hypothesis, I am
a. What is the probability of Type I error?
reducing the Type I error."
b. What is the probability of Type II error?
4. Which of the following statements are correct? Explain if
c. Discuss the consequences of Type I and Type II errors.
incorrect.
d. What is wrong with the statement,"! can prove that
a. "I accept the null hypothesis since sample eviden ce is
the individual is telling the truth on the basis of the
not inconsistent with the null hypothesis."
polygraph result."
b. "Since sample evidence cannot be supported by the null
10. The screening process for detecting a rare disease is not
hypothesis, I reject the null hypothesis."
perfect. Researchers have developed a blood test that is
c. "I can establish a given claim if sample evidence is
considered fairly reliable. It gives a positive reaction in 98%
consistent with the null hypothesis."
of the people w h o have that disease. However, it erroneously
d. "I cannot establish a given claim if the null hypothesis is
gives a positive reaction in 3% of the people w h o do not
not rejected."
have the disease. Answer the following questions using the
5. Construct the null and the alternative hypothese s for the null hypothesis as "the individual does not have the disease."
following tests:
a. What is the probability of Type I error?
a. Test if the mean w eight of cereal in a cereal box differs b. What is the probability of Type II error?
from 18 ounces.
c. Discuss the c onsequences o f Type I and Type II errors.
b. Test if the stock price increases on more than 60% o f the
d. What is wrong with the nurse's analysis,"The blood test
trading days.
result has proved that the individual is free of disease."
T E S T S T A T IS T IC F O R µ W H E N σ IS K N O W N
The value of the test statistic for the hypothesis test of the population mean µ
when the population standard deviation σ is known is computed as
where µ0 is the hypothesized mean value.
Note that the value of the test statistic z is evaluated at µ = µ0, which explains why we
need some form of the equality sign in the null hypothesis. Given that the population is
normally distributed with a known standard deviation, σ = 9, we compute the value of
the test statistic as Therefore, comparing with 67 is
identical to comparing z = 2.22 with 0, where 67 and 0 are the means of the unstandard
ized normal distribution and the standard normal distribution, respectively.
We now compute the p -value, which is the likelihood of obtaining a sample mean
that is at least as extreme as the one derived from the given sample, under the assumption
that the null hypothesis is true. Since in the above example we define the extreme
value as a sample mean of 71 or higher, and use the z table to compute the p -value as
Figure 9.1 shows the computed p -value.
Note that when the population mean equals 67, there is only a 1.32% chance that will be
71 or more. Therefore, if we decide to reject the null hypothesis, there is a 1.32% chance
that our decision will be erroneous. In order to make a decision, we need to determine if
the p -value is deemed small enough to reject the null hypothesis.
Remember that a Type I error occurs when we reject the null hypothesis when the null
hypothesis is true. The allowed probability of making a Type I error is α , which is also
referred to as the significance level of the test. We choose a value for α before implement
ing a hypothesis test. Most hypothesis tests are conducted using a significance level of 1,
5, or 10%, using α = 0.0 1 , 0.05, or 0.10, respectively. The decision rule is to reject the null
hypothesis when the p -value < α and not reject the null hypothesis when α -value ≥ α .
Suppose we choose α = 0.05 to conduct the above test. Therefore, we reject the null
hypothesis because 0.0132 < 0.05. This means that the sample data support the sociolo
gist’s claim that the average retirement age is greater than 67 years old. Individuals may
be working past the normal retirement age of 67 either because their savings have been
depleted due to the financial crisis and/or the fact that this generation is expected to out
live any previous generation and needs a job to pay the bills.
In the retirement age example of a right-tailed test, we calculated the p -value as
P(Z ≥ z). Analogously, for a left-tailed test the p -value is given by P (Z ≤ z). For a two-
tailed test, the extreme values exist on both sides of the distribution of the test statistic.
Given the symmetry of the z distribution, the p -value for a two-tailed test is twice that of
the p -value for a one-tailed test. It is calculated as 2P(Z ≥ z) if z > 0 or as 2P (Z ≤ z) if z < 0.
THE p - VALUE A P P R O A C H
Under the assumption that µ = µ 0, the p -value is the likelihood of observing a
sample mean that is at least as extreme as the one derived from the given sample. Its
calculation depends on the specification of the alternative hypothesis.
Alternative Hypothesis p-value
HA: µ > µ0 Right-tail probability: P(Z ≥ z )
HA: µ < µ0 Left-tail probability: P(Z ≤ z )
Figure 9.2a shows the p -value for a left-tailed test. Since the appropriate test statistic
follows the standard normal distribution, we calculate the p -value as P(Z ≤ z). When cal
culating the p -value for a right-tailed test (see Figure 9.2b), we find the area to the right
of the value of the test statistic z or equivalently, P(Z ≥ z). We should note here that if
z ≥ 0 for a left-tailed test or if z ≤ 0 for a right-tailed test, then H0 can never be rejected.
Figure 9.2c shows the p -value for a two-tailed test, calculated as 2P(Z ≤ z ) when z < 0
or as 2P(Z ≥ z) when z > 0.
We will now summarize the four-step procedure using the p -value approach.
SOLUTION:
a. Since we want to determine if the average is different from the predicted value
of $606.40, we specify the hypotheses as
H0: µ = 606.40
HA: µ ≠ 606.40
b. Note that is normally distributed since it is computed from a random sample
drawn from a normal population. Since σ is known, the test statistic follows
the standard normal distribution, and its value is
c. For a two-tailed test with a positive value for the test statistic, we compute
the p -value as 2P(Z ≥ 1.39). From the z table, we first find P(Z ≥ 1.39) =
1 – 0.9177 = 0.0823; so the p -value = 2 × 0.0823 = 0.1646. The decision
rule is to reject the null hypothesis if the p -value = 0.1646 is less than the cho
sen level of significance α .
d. We do not reject the null hypothesis since the p -value > α (0.1646 > 0.05).
Therefore, at the 5% significance level, we cannot conclude that average back-
to-school spending differs from $606.40 per family this year. The sample data
do not support the research analyst’s claim.
As shown in Figure 9.3, the decision rule is to reject H0 if z > 1.645. Since the value of the
test statistic, z = 2.22, exceeds the critical value, zα = 1.645, we reject the null hypothesis
and conclude that the mean age is significantly greater than 67. Thus, we confirm the
conclusion reached with the p -value approach.
We would like to stress that we always arrive at the same statistical conclusion whether
we use the p -value approach or the critical value approach. If z falls in the rejection re
gion, then the p -value must be less than α . Similarly, if z does not fall in the rejection
region, then the p -value must be greater than α . Figure 9.4 shows the equivalence of the
two results in the retirement age example of a right-tailed test.
We reject the null hypothesis because the p -value = 0.0132 is less than α = 0.05, or
equivalently, because z = 2.22 is greater than zα = 1.645.
The above example uses a right-tailed test to calculate the critical value as zα . Given
the symmetry of the z distribution around zero, the critical value for a left-tailed test is
– zα. For a two-tailed test, we split the significance level in half to determine two critical
values – zα/2 and zα/1 where P(Z ≥ zα/2) = α / 2 .
THE CRITICAL VALUE A P P R O A C H
The critical value approach specifies a region such that if the value of the test statis
tic falls into the region, the null hypothesis is rejected. The critical value depends on
the specification of the alternative hypothesis.
Figure 9.5 shows the three different scenarios of determining the critical value(s) depend
ing on the specification of the competing hypotheses.
Figure 9.5a shows a negative critical value for a left-tailed test where we reject the null
hypothesis if z < – z α . Similarly, Figure 9.5b shows a positive critical value for a right
tailed test where we reject the null hypothesis if z > z α. There are two critical values for
a two-tailed test, where we reject the null hypothesis when z < –zα/2 or when z > zα/2
(see Figure 9.5c).
We will now summarize the four-step procedure using the critical value approach.
When conducting a hypothesis test using the critical value approach, Steps 1 and 2
remain the same as with the p -value approach, that is,
Step 1. Specify the null and the alternative hypotheses.
Step 2. Specify the test statistic and compute its value.
Step 3. Find the critical value or values. Since the test statistic is assumed to
follow the z distribution, the critical value for a right-tailed test is zα , where
P(Z ≥ z α ) = α. Analogously, the critical value for a left-tailed test is – z α . The criti
cal values for a two-tailed test are – zα/2 and zα/2 , where P(Z ≥ zα/2) = α /2 .
Step 4. State the conclusion and interpret results. If the value of the test statistic
falls in the rejection region, the decision rule is to reject the null hypothesis. So for a
right-tailed test, we reject the null hypothesis if z > zα; for a left-tailed test we reject
the null hypothesis if z < – z α ; and for a two-tailed test, we reject the null hypothesis
if z > zα/2 or z < – z α / 2
EXAMPLE 9 .8
Repeat Example 9.7 using the critical value approach. Recall that a research analyst
wishes to determine if average back-to-school spending differs from $606.40.
A random sample of 30 households, drawn from a normally distributed population
with a standard deviation of $65, results in a sample mean of $622.85.
SOLUTION: Steps 1 and 2 remain the same as with the p -value approach. In
particular, the competing hypotheses are H0: µ = 606.40 versus HA: µ ≠ 606.40
and the value of the test statistic is For a two-
tailed test, we split the significance level in half to determine two critical values, one
on each side of the distribution of the test statistic. Given a 5% level of significance,
α /2 = 0.05/2 = 0.025 is used to derive zα/2 = z 0.025 as 1.96. Thus, the critical values
are – 1.96 and 1.96. As shown in Figure 9.6, the decision rule is to reject the H0 if
z > 1.96 or z < – 1.96.
Since z = 1.39 does not fall in the rejection region (– 1.96 < 1.39 < 1.96), we do
not reject the null hypothesis. At the 5% significance level, we cannot conclude that
average back-to-school spending differs from $606.40 per family. As always, our
conclusion is consistent with that using the p -value approach.
Confidence Intervals and Two-Tailed Hypothesis Tests
For a two-tailed hypothesis test, we can also construct a confidence interval to arrive at
the conclusion. Given that we conduct the hypothesis test at the α significance level, we
can use the sample data to determine a corresponding 100(1 – α )% confidence interval
for the population mean µ . If the confidence interval does not contain the value of the
hypothesized mean µ 0, then we reject the null hypothesis. If the confidence interval con
tains µ0, then we do not reject the null hypothesis.
EXAMPLE 9.9
Repeat Example 9.8 by constructing a confidence interval for µ .
resulting in the interval [599.59, 646.11], Since the hypothesized value of the popu
lation mean µ0 = 606.40 falls within the 95% confidence interval, we do not reject
H0. Thus, we arrive at the same conclusion as with the p -value and the critical value
approaches; that is, the sample data do not support the research analyst’s claim that
average back-to-school spending differs from $606.40 per family this year.
EXAMPLE 9 .10
A recent report in The New York Times (August 7, 2010) suggests that consumers
are spending less not only as a response to the economic downturn, but also due to a
realization that excessive spending does not make them happier. A researcher wants
to use debit card data to contradict the generally held view that the average amount
spent annually on a debit card is at least $8,000. She surveys 20 consumers and
asks them how much they spend annually on their debit cards. The results are given
below; the data, labeled Debit Spending, are also available on the text website.
7,960 7,700 7,727 7,704 8,543 7,661 7,767 8,761 7,530 8,128
7,938 7,771 7,272 8,113 7,727 7,697 7,690 8,000 8,079 7,547
Historical data show that the population standard deviation is $500 and that spending
on debit cards is normally distributed. Test the claim at a 1% level of significance.
SOLUTION: The researcher would like to establish that average spending on debit
cards is less than $8,000, or equivalently, µ < 8,000. Thus, we formulate the two
hypotheses as
H0: µ ≥ 8,000
HA: µ < 8,000
The normality condition of is satisfied since spending on debit cards is
normally distributed. Also, since the population standard deviation is known, the
test statistic is assumed to follow the z distribution. We know that the p -value for a
left-tailed test is given by P(Z ≤ z). The Excel function Z.TEST produces P(Z ≥ z),
which is the appropriate p -value for a right-tailed test. For a left-tailed test, as in this
example, we simply subtract the value that Excel returns from one. (For a two-tailed
test, if the value that Excel returns for P(Z ≥ z) is less than 0.50, we multiply this
value by 2 to obtain the p -value; if the value for P(Z ≥ z) is greater than 0.50, we
calculate the p -value as 2 × (1 – P(Z ≥ z)). We follow these steps.
A. Open the Debit Spending data file found on the text website.
B. Select Formulas > Insert Function > Z.TEST. This command returns the
p -value associated with a right-tailed test.
C. See Figure 9.7. We supply the following three arguments in the dialog box:
• Array is the data set. We select the 20 observations.
• X is the hypothesized mean under the null hypothesis. We enter 8000.
• Sigma is the value of the population standard deviation. We enter 500.
EXERCISES 9.2
When the population standard deviation σ is unknown, the test statistic for test
ing the population mean µ is assumed to follow the tdf distribution with n – 1
degrees of freedom, and its value is computed as
The next two examples show how we use the four steps outlined in the previous section
for hypothesis testing when we do not know the population standard deviation σ .
EXAMPLE 9.11
In the introductory case to this chapter, the dean at a large university in California
wonders if students at her university study less than the 1961 national average of
24 hours per week. She randomly selects 35 students and asks their average study
time per week (in hours). From their responses (see Table 9.1) she calculates a
sample mean of 16.37 hours and a sample standard deviation of 7.22 hours.
a. Specify the competing hypotheses to test the dean’s concern.
b. Calculate the value of the appropriate test statistic.
c. At a 5% significance level, specify the critical value(s) and the decision rule.
d. What is the conclusion to the hypothesis test?
SOLUTION:
a. This is an example of a one-tailed test where we would like to determine if the
mean hours studied is less than 24, or μ < 24. We formulate the competing
hypotheses as
H0: μ ≥ 24 hours
HA: μ < 24 hours
b. Recall that for any statistical inference regarding the population mean, it is
essential that the sample mean is normally distributed. This condition is
satisfied because the sample size is greater than 30, specifically n = 35.
Given we compute the value of the test statistic as
c. Since we have a left-tailed test, the critical value is given by −tα,df where
P(Tdf ≥ tα,df) = α . Referencing the t table with α = 0.05 and d f = n − 1 = 34,
we first find tα,df = t0.05,34= 1.691. Therefore, the critical value is −t0.05,34 =
– 1.691. As shown in Figure 9.8, the decision rule is to reject the null hypoth
esis if t34 < − 1.691.
d. We reject the null hypothesis since the value of the test statistic falls in the re
jection region (t34 = − 6.25 is less than − t0.05,34 = − 1.691). At the 5% signifi
cance level, average study time at the university is less than the 1961 average
of 24 hours per week.
EXAMPLE 9.12
As the introductory case to this chapter mentions, recent research finds that today’s
undergraduates study an average of 14 hours per week. Using the sample data from
Table 9.1, the dean would also like to test if the mean study time of students at her
university differs from today’s national average of 14 hours per week.
a. How should the dean formulate the competing hypotheses for this test?
b. Calculate the value of the test statistic.
c. Approximate the p -value.
d. At the 5% significance level, what is the conclusion to this test?
SOLUTION:
a. The dean would like to test if the mean study time of students at her university
differs from 14 hours per week. Therefore, we formulate the hypothesis for this
two-tailed test as
H0: μ = 14 hours
HA: μ ≠ 14 hours
c. Since t34 = 1.94 > 0, the p -value for a two-tailed test is 2P(T34 ≥ t34). Refer
encing the t table for d f = 34, we find that the exact probability P(T34 ≥ 1.94)
cannot be determined. Table 9.3 shows a portion of the t table where we see
that t34 = 1.94 lies between 1.691 and 2.032. This means that P(T34 ≥ 1.94)
will lie between P(T34 > 2.032) = 0.025 and P(T34 ≥ 1.691) = 0.05, that is
0.025 < P (T34 ≥ 1.94) < 0.05. Multiplying this probability by two results in a
p -value between 0.05 and 0.10, that is, 0.05 < p -value < 0.10. Shortly, we will
use Excel to find the exact p -value.
d. Since the p -value satisfies 0.05 < p -value < 0.10, it must be greater than
α = 0.05; we do not reject the null hypothesis. Therefore, the mean study time
of students at the university is not statistically different from today’s national
average of 14 hours per week.
Excel returns the exact p -value of 0.0607, so 2 × P ( T34 ≥ 1.94) = 0.0607. Consistent
with our conclusion in Example 9.12, we do not reject the null hypothesis since the
p -value of 0.0607 is greater than the significance level of α = 0.05.
S Y N O P S I S OF I N T R O D U C T O R Y C A S E
30. In order to conduct a hypothesis test of the population 36. The manager of a small convenience store does not want
mean, a random sample of 24 observations is drawn from a her customers standing in line for too long prior to a
normally distributed population. The resulting mean and the purchase. In particular, she is willing to hire an employee
standard deviation are calculated as 4.8 and 0.8, respectively. for another cash register if the average wait time of the
Use the p-value approach to conduct the following tests at customers is more than five minutes. She randomly
α = 0.05. observes the wait time (in minutes) of customers during
b. H0: μ = 4.5 against HA: μ ≠ 4.5 3.5 5.8 7.2 1.9 6.8 8.1 5.4
31. Use the critical value approach to conduct the same two a. Set up the null and the alternative hypotheses to
tests in the preceding question at α = 0.05. determine if the manager needs to hire another
32. In order to test if the population mean differs from 16, you employee.
draw a random sample of 32 observations and compute the b. Calculate the value o f the test statistic. What assumption
sample mean and the sample standard deviation as 15.2 and regarding the population is necessary to implement this
0.6, respectively. Use (a) the p-value approach and (b) the step?
c. Use the critical value approach to decide whether the 41. A mortgage specialist would like to analyze the average
manager needs to hire another employee at α = 0.10. mortgage rates for Atlanta, Georgia. He studies the following
d. Repeat the above analysis with the p-value approach. sample APR quotes. These are the annual percentage rates
(APR) for 30-year fixed loans. If he is willing to assume that
37. Small, energy-efficient, Internet-centric, new computers
these rates are randomly drawn from a normally distributed
are increasingly gaining popularity (New York Times, July
population, can he conclude that the mean mortgage rate
20,2008).These computers, often called netbooks, have
for the population exceeds 4.2%? Test the hypothesis at a
scant onboard memory and are intended largely for surfing
10% level of significance using (a) the p-value approach and
websites and checking e-mail. Some of the biggest companies
(b) the critical value approach.
are wary of the new breed of computers because their low
price could threaten PC makers'already thin profit margins. Financial Institution APR
An analyst comments that the larger companies have a cause G Squared Financial 4.125%
for concern since the mean price of these small computers has
Best Possible Mortgage 4.250
fallen below $350. She examines six popular brands of these
Hersch Financial Group 4.250
small computers and records their retail price as:
Total Mortgages Services 4.375
$322 $269 $373 $412 $299 $389
Wells Fargo 4.375
a. What assumption regarding the distribution of the price Quicken Loans 4.500
of small computers is necessary to test the analyst's
Amerisave 4.750
claim?
b. Specify the appropriate null and the alternative Source: MSN Money.com; data retrieved October 1, 2010.
hypotheses to test the analyst's claim. 42. In September 2007, U.S. home prices continued to fall at a
c. Calculate the value of the test statistic. record pace, and price declines in Los Angeles and Orange
d. At the 5% significance level, what is the critical value(s)? counties in California outpaced other major metropolitan
What is the conclusion to the test? Should the larger areas (Los Angeles Times, November 28,2007). The report
computer companies be concerned? was based on the Standard & Poor's/Case-Shiller index
38. A local brewery wishes to ensure that an average of that measures the value of single-family homes based on
12 ounces of beer is used to fill each bottle. In order to their sales histories. According to this index, the prices
analyze the accuracy of the bottling process, the bottler in San Diego dropped by an average of 9.6% from a year
takes a random sample of 48 bottles. The sample mean earlier. Assume that the survey was based on recent sales
weight and the sample standard deviation of the bottles are of 34 houses in San Diego that also resulted in a standard
11.80 ounces and 0.8 ounce, respectively. deviation of 5.2%. Can we conclude that the mean drop
of all home prices in San Diego is greater than the 7%
a. State the null and the alternative hypotheses for the test.
drop in Los Angeles? Use a 1% level of significance for the
b. Do you need to make any assumption regarding the
analysis.
population for testing?
43. (Use Excel) One of the consequences of the economic
c. At α = 0.05 what is the critical value(s)? What is the
meltdown has been a free fall of the stock market's average
decision rule?
price/earnings ratio, or P/E ratio (The Wall Street Journal,
d. Make a recommendation to the bottler.
August 30,2010). Generally, a high P/E ratio suggests that
39. Based on the average predictions of 47 members of the investors are expecting higher earnings growth in the future
National Association of Business Economists (NABE), the U.S. compared to companies with a lower P/E ratio. An analyst
gross domestic product (GDP) will expand by 3.2% in 2011 wants to determine if the P/E ratio of firms in the footwear
(The Wall Street Journal, May 23,2010). Suppose the sample
industry is different from the overall average of 14.9. The
standard deviation of their predictions was 1%. At a 5% table below shows the P/E ratios for a sample of seven firms
significance level, test if the mean forecast GDP of all NABE in the footwear industry:
members is greater than 3%.
Firm P/E Ratio
40. A car manufacturer is trying to develop a new sports car.
Brown Shoe Co., Inc. 20.54
Engineers are hoping that the average amount of time that
the car takes to go from 0 to 60 miles per hour is below Collective Brands, Inc. 9.33
6 seconds. The car company tested 12 of the cars and clocked Crocs, Inc. 22.63
their performance times. Three of the cars clocked in at DSW, Inc. 14.42
5.8 seconds, 5 cars at 5.9 seconds, 3 cars at 6.0 seconds, Nike, Inc. 18.68
and 1 car at 6.1 seconds. At a 5% level of significance, test
Skechers USA, Inc. 9.35
if the new sports car is meeting its goal to go from 0 to
Timberland Co. 14.93
60 miles per hour in less than 6 seconds. Assume a normal
distribution for the analysis. Source: http://biz.yahoo.com ; data retrieved August 23, 2010.
a. State the null and the alternative hypotheses in order The complete data set is on the text website and is labeled
to test whether the P/E ratio of firms in the footwear Debt Payments.
industry differs from the overall average of 14.9. City Debt Payments
b. What assumption regarding the population is necessary? Washington, D.C. $1,285
c. Use Excel to calculate descriptive statistics for the data, Seattle 1,135
then use the function T.DIST.2T to calculate the p-value. ⋮ ⋮
d. At α = 0.10 what is the conclusion?
Pittsburgh 763
44. (Use Excel) Access the miles per gallon (MPG) data on
Source: www.Experian.com, November 11, 2010.
the text website labeled MPG.
a. State the null and the alternative hypotheses in order to a. State the null and the alternative hypotheses in order
test w hether the average MPG differs from 95. to test whether average monthly debt payments are
greater than $900.
b. Use Excel to calculate descriptive statistics for the data,
then use the function T.DIST.2T to calculate the p-value. b. What assumption regarding the population is necessary
to implement this step?
c. At α = 0.05 can you conclude that the MPG differs from
c. Use Excel to calculate descriptive statistics for the data.
95?
Calculate the value of the test statistic.
45. (Use Excel) A recent study fo u n d th a t consum ers are
d. Use Excel's function T.DIST.RT to calculate the p-value.
m aking average m o n th ly d e b t paym ents o f $983 (Experian.
com , N ove m b e r 1 1 , 2 0 1 0). The a ccom panying ta b le shows a
e. At α = 0.05 what is your conclusion?
p o rtio n o f average d e b t paym ents fo r 26 m e tro p o lita n areas.
TEST ST A T IS T IC FOR p
The test statistic for the hypothesis test of the population proportion p is assumed
to follow the z distribution and its value is computed as where
and p0 is the hypothesized value of the population proportion.
The following examples elaborate on the four-step procedure for a hypothesis test of the
population proportion.
EXAMPLE 9.13
A popular weekly magazine asserts that fewer than 40% of households in the United
States have changed their lifestyles because of escalating gas prices. A recent sur
vey of 180 households finds that 67 households have made lifestyle changes due to
escalating gas prices.
a. Specify the competing hypotheses to test the magazine’s claim.
b. Calculate the value of the test statistic and the corresponding p -value.
c. At a 10% level of significance, what is the conclusion?
SOLUTION:
a. We wish to establish that the population proportion is less than 0.40, or
p < 0.40. Thus, we construct the competing hypotheses as
H0: p ≥ 0.40
HA: p < 0.40
b. We first ensure that we can use the normal distribution approximation for the
test. Since both np0 and n( 1 − p 0) exceed 5, the normal approximation is jus
tified. We use the sample proportion. to compute the
value of the test statistic as
EXAMPLE 9.14
Nearly one in three children and teens in the U.S. are obese or overweight (Health,
October 2010). A health practitioner in the Midwest collects data on 200 children
and teens and finds that 84 of them are either obese or overweight.
a. The health practitioner believes that the proportion of obese and overweight
children in the Midwest is not representative of the national proportion. Specify
the competing hypotheses to test her claim.
b. Calculate the value of the test statistic.
c. At the 1% significance level, specify the critical value(s) and the decision rule.
d . Do the sample data support the health practitioner’s belief?
SOLUTION:
a. The parameter of interest is again the population proportion p. The health
practitioner wants to test if the population proportion of obese or overweight
children in the Midwest differs from the national proportion of 1/3 = 0.33. We
construct the hypotheses as
H0: p = 0.33
HA: p ≠ 0.33
d. Since the value of the test statistic, z = 2.71, is greater than 2.575, the
appropriate decision is to reject the null hypothesis. Therefore, at the 1%
significance level, the practitioner concludes that the proportion of obese or
overweight children in the Midwest is not the same as the national proportion
of 0.33. Given that the test statistic fell in the right side of the distribution, the
practitioner can conduct further analysis to determine whether or not the pro
portion of obese or overweight children in the Midwest is significantly greater
than the national proportion. If this is the case, the obesity epidemic may be
more problematic in the Midwest than elsewhere and may require more educa
tional initiatives to curb its harmful ramifications.
EXERCISES 9.4
Concepts Applications
46. Consider the following hypotheses: 55. A recent study by Allstate Insurance Co. finds that 82% of
H0: p ≥ 0.38 teenagers have used cell phones while driving (The Wall
Street Journal, May 5 , 2010). In O cto b er 2010, Massachusetts
HA: p < 0.38
e nacted a law th a t fo rb id s cell p h o ne use by drivers under
Compute the p-value based on the following sample th e age o f 18. A p o licy analyst w o u ld like to d e term in e
information. w h e th e r th e law has decreased th e p ro p o rtio n o f drivers
un d er th e age o f 18 w h o use a cell phone.
a. State the null and the alternative hypotheses to test the
policy analyst's objective.
b. Suppose a sample of 200 drivers under the age of
47. Which sample information in the preceding question enables 18 results in 150 who still use a cell phone while driving.
us to reject the null hypothesis at α = 0.01 and at α = 0.10? What is the value of the test statistic? What is the
p-value?
48. Consider the following hypotheses:
c. At α = 0.05 has the law been effective?
H0: p = 0.32
d. Repeat this exercise using the critical value approach
HA: p ≠ 0.32 with α = 0.05.
Compute the p-value based on the following sample 56. Due to the recent economic downturn, Americans have
information started raiding their already fragile retirement accounts to
endure financial hardships such as unemployment, medical
emergencies, and buying a home (MSN Money, July 16,
2008). It is reported that between 1998 and 2004, about
12% of families with 401 (k) plans borrowed from them. An
economist is concerned that this percentage now exceeds
.
20%. He randomly surveys 190 households with 401 (k) plans
and finds that 50 are borrowing against them.
50. Determine the critical value or values for the following tests
a. Set up the null and the alternative hypotheses to test the
of the population proportion. The analysis is conducted at a
economist's concern.
5% level of significance.
b. Compute the value of the appropriate test statistic.
a. H0: p ≤ 0.22; HA: p > 0.22
c. Use the p-value approach to test if the economist's
b. H0: p = 0.69; HA: p + 0.69
concern is justifiable a t α = 0.05.
c. H0: p ≥ 0.56; HA: p < 0.56
57. The margarita is one of the most common tequila-based
51. In order to conduct a hypothesis test of the population cocktails, made with tequila mixed with triple sec and lime
proportion, you sample 320 observations that result in or lemon juice, often served with salt on the glass rim.
128 successes. Use the p-value approach to conduct the A common ratio for a margarita is 2:1:1, which includes
following tests at α = 0.05. 50% tequila, 25% Triple Sec, and 25% fresh lime or lemon
a. H0: p ≥ 0.45; HA: p < 0.45 juice. A manager at a local bar is concerned that the
b. H0 p = 0.45; HA: p ≠ 0.45 bartender does not use the correct proportions in more than
52. Repeat the preceding exercise using the critical value 50% of margaritas. He secretly observes the bartender and
approach at α = 0.01. finds that he used the correct proportions in only 10 out of
30 margaritas. Use the critical value approach to test if the
53. You would like to determine if the population probability
manager's suspicion is justified at α = 0.05.
of success differs from 0.70. You find 62 successes in
80 binomial trials. Implement the test at a 1% level of 58. A movie production company is releasing a movie with the
significance. hopes of many viewers returning to see the movie in the
theater for a second time. Their target is to have 30 million
54. You would like to determine if more than 50% of the
viewers, and they want more than 30% of the viewers to
observations in a population are below 10. At α = 0.05,
return to see the movie again. They show the movie to a
conduct the test on the basis of the following 20 sample
test audience of 200 people, and after the movie they asked
observations:
them if they would see the movie in theaters again. Of the
8 12 5 9 14 11 9 3 7 8 test audience, 68 people said they would see the movie
12 6 8 9 2 6 11 4 13 10 again.
a. At a 5% level o f significance, test if more than 30% o f the 61. New research shows that many banks are unw ittingly training
viewers w ill return to see the movie again. their online customers to take risks with their passwords
b. Repeat the analysis at a 10% level o f significance. and other sensitive account information, leaving them more
c. Interpret your results. vulnerable to fraud (Yahoo.com, July 23, 2008). Even web-
savvy surfers could find themselves the victims of identity theft
59. Recent research commissioned by Vodafone suggests that
because they have been conditioned to ignore potential signs
older workers are the happiest employees (BBC News, July
about whether the banking site they are visiting is real or a
2 1 , 2008). The report documents tha t 70% o f older workers
bogus site served up by hackers. Researchers at the University
in England feel fulfilled, compared w ith just 50% o f younger
of Michigan found design flaws in 78% o f the 214 U.S. financial
workers. A dem ographer believes tha t an identical pattern
institution websites they studied. Is the above sample evidence
does not exist in Asia. A survey o f 120 older workers in Asia
sufficient to conclude that more than three out of four financial
finds tha t 75 feel fulfilled. A similar survey finds that 58% o f
institutions that offer online banking facilities are prone to
210 younger workers feel fulfilled.
fraud? Use a 5% significance level for the test.
a. At a 5% level o f significance, test if older workers in Asia
62. The Social Security Administration is not expected to provide
feel less fulfilled than the ir British counterparts.
any increases in Social Security benefits for the second
b. At a 5% level o f significance, test if younger workers in
straight year (US News & World Report, October 4 , 2010).
Asia feel more fulfilled than the ir British counterparts.
With increasing medical prices, it is claimed that more than
60. A politician claims that he is supported by a clear m ajority 60% of seniors are likely to make serious adjustments to
of voters. In a recent survey, 24 out o f 40 randomly selected their lifestyle. Test this claim at a 1% level of significance if
voters indicated that they would vote for the politician. Is the in a survey of 140 seniors, 90 reported that they have made
politician's claim justifiable at a 5% level o f significance? serious adjustments to their lifestyle.
Trevor Jones is a newspaper reporter who is interested in verifying the concerns of the
local politician.
Trevor wants to use the sample information to:
1. Determine if the mean income of Latino households has fallen below the 2008 level
of $49,000.
2. Determine if the proportion of Latino households making less than $30,000 has risen
above 2 0 %.
One of the hotly debated topics in the United States is that of growing income inequal
ity. Market forces such as increased trade and technological advances have made highly
Sample
skilled and well-educated workers more productive, thus increasing their pay. Institu
tional forces, such as deregulation, the decline of unions, and the stagnation of the mini
Report—
mum wage, have contributed to income inequality. Arguably, this income inequality has
been felt by minorities, especially African-Americans and Latinos, since a very high
Income
proportion of both groups is working class. The condition has been further exacerbated
by the recent economic meltdown.
Inequality in
A sample of 36 Latino households resulted in a mean household income of $46,278
with a standard deviation of $19,524. The sample mean is below the 2008 level of $49,000.
the United
In addition, nine Latino households, or 25%, make less than $30,000; the corresponding
percentage in 2008 was 20%. Based on these results, a politician concludes that current
States
market conditions continue to negatively impact the welfare of Latinos. However, it is
essential to provide statistically significant evidence to substantiate these claims. Toward
this end, formal tests of hypotheses regarding the population mean and the population
proportion are conducted. The results of the tests are summarized in Table 9.A.
TABLE 9.A Test Statistic Values and p-Values for Hypothesis Tests
When testing whether the mean income of Latino households has fallen below the 2008
level of $49,000, a test statistic value of – 0.84 is obtained. Given a p -value of 0.2033, the
null hypothesis regarding the population mean, specified in Table 9.A, cannot be rejected
at any reasonable level of significance. Similarly, given a p -value of 0.2266, the null
hypothesis regarding the population proportion cannot be rejected. Therefore, sample evi
dence does not support the claims that the mean income of Latino households has fallen
below $49,000 or that the proportion of Latino households making less than $30,000 has
risen above 20%. Perhaps the politician’s remarks were based on a cursory look at the
sample statistics and not on a thorough statistical analysis of the incomes.
Conceptual Review
LO 9.1 Define the null hypothesis and the alternative hypothesis.
Every hypothesis test contains two competing hypotheses: the null hypothesis, de
noted H0, and the alternative hypothesis, denoted HA. We can think of the null hypoth
esis as corresponding to a presumed default state of nature or status quo, whereas the
alternative hypothesis contradicts the default state or status quo.
On the basis of sample information, we either reject H0 or do not reject H0.
As a general guideline, whatever we wish to establish is placed in the alternative hy
pothesis. If we reject the null hypothesis, we are able to conclude that the alternative
hypothesis is true.
Hypothesis tests can be one-tailed or two-tailed. A one-tailed test allows the rejection
of the null hypothesis only on one side of the hypothesized value of the population
parameter. In a two-tailed test, the null hypothesis can be rejected on both sides of the
hypothesized value of the population parameter.
LO 9.2 Distinguish between Type I and Type II errors.
Since the statistical conclusion of a hypothesis test relies on sample data, there are two
types of errors that may occur: a Type I error or a Type II error. A Type I error is com
mitted when we reject the null hypothesis when it is actually true. On the other hand, a
Type II error is made when we do not reject the null hypothesis when it is actually false.
W e d e n o te th e p ro b a b ility o f a T y p e I e rro r b y α a n d th e p ro b a b ility o f a T ype II e rro r b y β .
For a given sample size n, a decrease (increase) in α will increase (decrease) β . However,
both α and β will decrease if the sample size n increases.
LO 9.3 Explain the steps of a hypothesis test using the p -value approach.
Every hypothesis test can be implemented by following a four-step procedure. There
are two equivalent approaches, namely the p -value approach and the critical value
approach. For the p -value approach, we follow these four steps:
Step 1. Specify the null and the alternative hypotheses. We identify the relevant
population parameter of interest, determine whether it is a one- or a two-tailed test and,
most importantly, include some form of the equality sign in the null hypothesis and
place whatever we wish to establish in the alternative hypothesis.
Step 2. Specify the test statistic and compute its value. In this chapter, we derive
the value of the test statistic by converting the estimate of the relevant population
parameter into its corresponding standardized value, either z or tdf.
Step 3. Calculate the p -value. We find the probability that the test statistic is as ex
treme as its value computed from the given sample. If the test statistic is a standard
normal random variable Z, then the p -value is calculated as
• P(Z ≥ z) for a right-tailed test,
• P(Z ≤ z) for a left-tailed test, or
• 2P(Z ≥ z) if z > 0 or 2P (Z ≤ z) if z < 0 for a two-tailed test.
Z and z are replaced by Tdf and tdf if the assumed test statistic follows the tdf distribu
tion with n – 1 degrees of freedom.
Step 4. State the conclusion and interpret results. The decision rule is to reject the
null hypothesis if the p -value < α , where α is the chosen significance level.
LO 9.4 Explain the steps of a hypothesis test using the critical value approach.
For the critical value approach, we follow these four steps:
Step 1 and Step 2 are the same as the p -value approach.
S te p 3 . F in d th e critic a l v a lu e o r v a lu es. T h e critical v alu e(s) d efin es th e reg io n o f v alu es
o f th e te st statistic fo r w h ic h th e n u ll h y p o th e s is c a n b e rejected . F o r a g iv e n α , th e c ritical
v a lu e (s) is fo u n d as:
CASE STUDIES
70 90 80 100 20 70 90
80 30 80 90 90 80 30
10
c H A P T
LEARNING OBJECTIVES
E R
Concerning Two
Populations
After reading LO 10.1 Make inferences about the difference between two population means based
this chapter on independent sampling.
you should
be able to: L O 1 0 .1 Make inferences about the mean difference based on matched-pairs
sampling.
LO 10.3 Make inferences about the difference between two population proportions
based on independent sampling.
In the preceding tw o chapters, we used estimation and hypothesis testing to analyze a single pa
rameter, such as the population mean and the population proportion. In this chapter we extend
our discussion from the analysis o f a single population to the comparison o f tw o populations. We
first analyze differences between tw o population means. For instance, an economist may be inter
ested in analyzing the salary difference between male and female employees. Similarly, a marketing
researcher m ig h t w a nt to compare the operating lives o f tw o popular brands o f batteries. In these
examples, we use independent sampling for the analysis. We will also consider the mean difference
o f tw o populations based on matched-pairs sampling. An example o f such a case w ould be a con
sumer group activist w anting to analyze the mean w e ig h t o f customers before and after they enroll
in a new diet program. Finally, we look at q ualitative data and compare the difference between tw o
population proportions. For instance, marketing executives and advertisers are often interested in
the different preferences between males and females when determ ining where to target advertising
dollars. In each o f the statistical inferences concerning tw o populations, we first develop the proce
dure for estimation and then follow w ith hypothesis testing.
I N T R O D U C T O R Y CASE
TABLE 10.1 Average Caloric Intake Before and After Menu-Labeling Ordinance
IN D E P E N D E N T R A N D O M SAMPLES
Two (or more) random samples are considered independent if the process that gen
erates one sample is completely separate from the process that generates the other
sample. The samples are clearly delineated.
C O N F I D E N C E I N T E R V A L F OR μ 1 − μ 2
A 100( 1 − α )% confidence interval of the difference between two population means
μ1 − μ2 is given by:
Note that in the case when we construct a confidence interval for μ1 − μ2 where σ12 and σ22
are unknown but assumed equal, we calculate a pooled estimate of the common variance
Sp2. In other words, because the two populations are assumed to have the same population
variance, the two sample variances S12 and S22 are simply two separate estimates of this
population variance. We estimate the population variance by a weighted average of S12and
S22, where the weights applied are their respective degrees of freedom relative to the total
number of degrees of freedom. In the case when σ12 and σ22 are unknown and cannot be
assumed equal, we cannot calculate a pooled estimate of the population variance because
of different variabilities in the two populations.
EXAMPLE 10.1
A consumer advocate analyzes the nicotine content in two brands of cigarettes.
A sample of 20 cigarettes of Brand A resulted in an average nicotine content of
1.68 milligrams with a standard deviation of 0.22 milligram; 25 cigarettes of Brand
B yielded an average nicotine content of 1.95 milligrams with a standard deviation
of 0.24 milligram.
Brand A Brand B
s1 = 0.22 mg s2 = 0.24 mg
n1 = 20 n 2 = 25
Construct a 95% confidence interval for the difference between the two popula
tion means. Nicotine content is assumed to be normally distributed. In addition, the
population variances are unknown but assumed equal.
In other words, the 95% confidence interval for the difference between the two
means ranges from − 0.41 to − 0.13. Shortly, we will see how to use this interval to
conduct a two-tailed hypothesis test.
H0 μ1 − μ2 = d0 H0 μ1 − μ2 ≤ d0 H0 μ1 − μ2 ≥ d0
HA:μ1 − μ2 ≠ d0 HA:μ1− μ2 > d0 HA:μ1 − μ2 < d0
In most applications, the hypothesized difference d0 between two population means μ1,
and μ 2 is zero. In this scenario, a two-tailed test determines whether the two means differ
from one another; a right-tailed test determines whether μ1 is greater than μ2; and a left
tailed test determines whether μ1 is less than μ 2.
We can also construct hypotheses where the hypothesized difference d0 is a value
other than zero. For example, if we wish to determine if the mean return of an emerging
market fund is more than 2% higher than that of a developed market fund, the resulting
hypotheses are H0:μ 1 − μ 2 ≤ 2 versus HA:μ1 − μ2 ≤ 2.
EXAMPLE 10.2
Revisit Example 10.1.
a. Specify the competing hypotheses in order to determine whether the average
nicotine levels differ between Brand A and Brand B.
b. Using the 95% confidence interval, what is the conclusion to the test?
SOLUTION:
a. We want to determine if the average nicotine levels differ between the two
brands, or μ1≠μ2, so we formulate a two-tailed hypothesis test as
H0: μ1 − μ2 = 0
HA:μ 1 − μ 2 ≠ 0
b. In Example 10.1, we calculated the 95% confidence interval for the differ
ence between the two means as − 0.27 ± 0.14; or equivalently, the confidence
interval ranges from − 0.41 to − 0.13. This interval does not contain zero, the
value hypothesized under the null hypothesis. This information allows us to
reject H0; the sample data support the conclusion that average nicotine levels
between the two brands differ at the 5% significance level.
TEST S TA T IS T IC FOR T E S T IN G μ 1 −
μ2
1. If σ12 and σ22 are known, then the test statistic is assumed to follow the z distri
bution and its value is calculated as
2. If σ12 and σ22 are unknown but assumed equal, then the test statistic is assumed
to follow the tdf distribution and its value is calculated as
where a n d d f = n 1 + n 2 − 2.
3. If σ12 and σ22 are unknown and cannot be assumed equal, then the test
statistic is assumed to follow the tdf distribution and its value is calculated
is rounded
We should note that these tests are valid only when the sampling distribution of
is normal.
EXAMPLE 10.3
An economist claims that average weekly food expenditure of households in City 1
is more than that of households in City 2. She surveys 35 households in City 1 and
obtains an average weekly food expenditure of $164. A sample of 30 households in
City 2 yields an average weekly food expenditure of $159. Historical data reveals
that the population standard deviation for City 1 and City 2 are $12.50 and $9.25,
respectively.
City 1 City 2
σ1 = $12.50 σ 2 = $9.25
n1= 35 n2 = 30
EXAMPLE 10.4
Table 10.2 shows annual return data for 10 firms in the gold industry and 10 firms
in the oil industry; these data, labeled Gold and Oil, are also available on the text
website. Can we conclude at the 5% significance level that the average returns in the
two industries differ? Here we assume that we are sampling from two normal popu
lations and that the population variances are unknown and not equal. The assump
tion concerning the population variances is reasonable since variance is a common
measure of risk; we cannot assume that the risk from investing in the gold industry
is the same as the risk from investing in the oil industry.
TABLE 10.2 Annual Returns (in percent)
Gold Oil
6 −3
15 15
19 28
26 18
2 32
16 31
31 15
14 12
15 10
16 15
SOLUTION: We let μ1 denote the mean return for the gold industry and μ2 denote
the mean return for the oil industry. Since we wish to test whether the mean returns
differ, we set up the null and alternative hypotheses as
H0: u1 − μ2 = 0
HA: μ1 − μ2 ≠ 0
Given that we are testing the difference between two means when the popula
tion variances are unknown and not equal, we need to calculate
Recall that the calculation for the degrees of freedom for the corresponding test
statistic is rather involved. Using one command on Excel, we are provided with the
value of the test statistic, the degrees of freedom, and the p -value, as well as the
relevant critical values. We follow these steps.
a. Open the Gold and Oil data found on the text website.
b. Choose Data > Data Analysis > t-Test: Two-Sample Assuming Unequal
Variances > OK. (Note: Excel provides two other options when we want
to test the difference between two population means and we have access to
the raw data. If the population variances are known, we can use the option
z-Test: Two-Sample for Means. If the population variances are unknown but
assumed equal, we can use the option t-Test: Two-Sample Assuming Equal
Variances.)
c. See Figure 10.1. In the t-Test: Two-Sample Assuming Unequal Variances dia
log box, choose Variable 1 Range and select the gold data. Then, choose Vari
able 2 Range and select the oil data. Enter a Hypothesized Mean Difference of
0 since d0 = 0, check the Labels box if you include Gold and Oil as headings,
and enter an α value of 0.05 since the test is conducted at the 5% significance
level. Choose an output range and click OK.
We now show that we reach the same conclusion concerning the mean return of
these two industries using the critical value approach as well as the confidence
interval approach. Given α = 0.05, the relevant critical values for this two-tailed
test are − 2.1098 and 2.1098 (see t Critical two-tail in Table 10.3). The decision
rule is to reject H0 if t17 > 2.1098 or t17 < − 2.1098. The value of the test statistic
is t17 = − 0.3023 (see t Stat in Table 10.3). Since the value of the test statistic is
between the two critical values, − 2.1098 < − 0.3023 < 2.1098, we cannot reject
the null hypothesis. As always, our conclusion is consistent with that of the p -value
approach.
Finally, given the information in Table 10.3, it is also possible to calculate the
corresponding 95% confidence interval for μ1 – μ2 as
That is, a 95% confidence interval for the difference between the two means ranges
from − 10.37 to 7.77. We note that this interval contains zero, the value hypoth
esized under the null hypothesis. Using a 95% confidence interval, we again see
that the sample data support the conclusion that the population mean returns do not
differ at the 5% significance level.
Caution
In Example 10.4, we may have made a strong assumption that the populations were nor
mally distributed. We could not invoke the central limit theorem as we had small sample
sizes. In Chapter 12, we will explore tests that check for normality. If we wish to draw
inferences about μ1 − μ2 from nonnormal populations, we need to apply nonparametric
techniques such as the Wilcoxon rank-sum test for independent samples, a test discussed
in Chapter 20.
E X E R C I S E S 10.1
Sample 1 Sample 2
12.1 8.9
a. Using the p-value approach, test whether the population 9.5 10.9
means differ at the 5% significance level. 7.3 11.2
b. Repeat the hypothesis test using the critical value 10.2 10.6
approach.
8.9 9.8
4. Consider the following competing hypotheses and 9.8 9.8
accompanying sample data. The two populations are known
7.2 11.2
to be normally distributed.
10.2 12.1
a. Implement the test at the 5% level under the c. What is the inference of the test at a 10% level of
120 110
a. Implement the test at the 5% level under the a. Construct the relevant hypothesis to test if the means of
assumption that the population variances are unknown the two populations differ.
a. Set up the hypotheses to test if the report's conclusion critical value approach.
also applies to Lucille's university. 14. In May 2008, CNN reported that Sports Utility Vehicles (SUVs)
b. Calculate the value of the test statistic and its p-value. are plunging toward the "endangered" list. Due to soaring
c. At the 5% significance level, can we conclude that there oil prices and environmental concerns, consumers are
is a "community college penalty" at Lucille's university? replacing gas-guzzling vehicles with fuel-efficient smaller
cars. As a result, there has been a big drop in the demand
11. The Chartered Financial Analyst (CFA®) designation is
for new as well as used SUVs. A sales manager of a used car
fast becoming a requirement for serious investment
dealership believes that it takes 30 days longer to sell an SUV
professionals. It is an attractive alternative to getting an
as compared to a small car. In the last two months, he sold 18
MBA to students wanting a career in investment. A student
SUVs that took an average of 95 days to sell with a standard
of finance is curious to know if a CFA designation is a
deviation of 32 days. He also sold 38 small cars with an
more lucrative option than an MBA. He collects data on 38
average of 48 days to sell and a standard deviation of 24 days.
recent CFAs with a mean salary of $138,000 and a standard
a. Construct the null and the alternative hypotheses to
deviation of $34,000. A sample of 80 MBAs results in a mean
contradict the manager's claim.
salary of $130,000 with a standard deviation of $46,000.
b. Compute the value of the test statistic under the
a. Use the p-value approach to test if a CFA designation is
assumption that the variability of selling time for the
more lucrative than an MBA at the 5% significance level.
SUVs and the small cars is the same.
Do not assume that the population variances are equal.
Make sure to state the competing hypotheses. c. Implement the test at α = 0.10 and interpret your
results.
b. Repeat the analysis with the critical value approach.
15. (Use Excel) A consumer advocate researches the length 17. (Use Excel) Baseball has always been a favorite pastime in
o f life between tw o brands of refrigerators, Brand A and America, and is rife w ith statistics and theories. W hile baseball
Brand B. He collects data on the longevity o f 40 refrigerators purists may disagree, to an applied statistician no to p ic in
for Brand A and repeats the sampling for Brand B. These data baseball is to o small or hypothesis to o unlikely. In a recent
are measured in years and can be found on the text website, paper, researchers at Wayne State University showed th a t
labeled Refrigerator Longevity. m ajor-league players w h o have nicknames live 21/2 years longer
a. Specify the competing hypotheses to test whether the than those w ith o u t them (The Wall Street Journal, July 1 6 , 2009).
average length o f life differs between the tw o brands. You do n o t believe in this result and decide to conduct a test to
b. Using the appropriate commands in Excel, find the value prove it w rong. Consider th e fo llo w in g data on th e lifespan o f a
of the test statistic. Assume th at σA2 = 4.4 and σB2 = 5.2. player and th e nickname variable, w hich equals 1 if the player
What is the p -value? had a nicknam e and 0 oth erw ise . These data can also be found
on th e te x t website, labeled Nicknames.
c. At the 5% significance level, w hat is the conclusion?
Years Nickname Years Nickname Years Nickname
16. (Use Excel) According to a study published in the New
England Journal o f Medicine, overweight people on low- 74 1 61 0 68 1
carbohydrate and Mediterranean diets lost more weight 62 1 64 0 68 0
and got greater cardiovascular benefits than people on a 67 1 70 0 64 1
conventional low-fat diet (The Boston Globe, July 17, 2008). 73 1 71 1 67 1
A nutritionist wishes to verify these results and follows 30
49 1 69 1 64 0
dieters on the low-carbohydrate and Mediterranean diets
62 0 56 0 63 1
and 30 dieters on the low-fat diet. These data (measured in
pounds) can be found on the text website, labeled Different
56 0 68 1 68 1
Diets. 63 0 70 1 68 1
a. Set up the hypotheses to test the claim that the 80 1 79 1 74 0
mean w eight loss for those on low-carbohydrate or 65 1 67 0 64 0
Mediterranean diets is greater than the mean w eight
a. Create tw o subsamples consisting o f players w ith and
loss for those on a conventional low-fat diet.
w ith o u t nicknames. Calculate the average longevity for
b. Using the appropriate commands in Excel, find the
each subsample.
value of the test statistic. Assume that the population
b. Specify the hypotheses to contradict the claim made by
variances are equal and that the test is conducted at the
researchers at Wayne State University.
5% significance level. What are the critical value(s) and
c. State the conclusion o f the test using a 5% level o f
the rejection rule?
significance. Assume the population variances are
c. At the 5% significance level, can the nutritionist
unknown but equal.
conclude that overweight people on low-carbohydrate
d. What assumptions did you make regarding the
or Mediterranean diets lost more w eight than people on
populations?
a conventional low-fat diet?
When constructing a confidence interval for the mean difference μ D, we follow the same
general format of point estimate ± margin of error.
C O N F I D E N C E I N T E R V A L FOR μ D
A 100(1 – a )% confidence interval of the mean difference μ D is given by:
where and sD are the mean and the standard deviation, respectively, of the n sam
ple differences, and d f = n – 1.
In the next example, the values for and sD are explicitly given; we will outline the cal
culations when we discuss hypothesis testing.
EXAMPLE 10.5
A manager is interested in improving productivity at a plant by changing the lay
out of the workstation. She measures the productivity of 10 workers before the
change and again after the change. She calculates the following summary statistics:
and n = 10. Construct a 95% confidence interval for the mean
difference.
SOLUTION: In order to construct a 95% confidence interval for the mean difference,
we use and a = 0.05, we find
2α/2.df = t0.025.9 = 2.262. Plugging the relevant values into the formula, we calculate
That is, a 95% confidence interval for the
mean difference ranges from 0.36 to 16.64. This represents a fairly wide interval,
caused by the high standard deviation sDof the 10 sample differences.
Hypothesis Test for μ D
As before, we generally want to test whether the mean difference μ D is equal to, greater
than, or less than a given hypothesized mean difference d0, or:
EXAMPLE 10.6
Using the information from Example 10.5, can the manager conclude at the 5%
significance level that there has been a change in productivity since the adoption of
the new workstation?
SOLUTION: In order to determine whether or not there has been a change in the
mean difference, we formulate the null and the alternative hypotheses as
H0: μD = 0
Ha: μ D ≠ 0
In Example 10.5, we found that a 95% confidence interval for the mean difference
ranges from 0.36 to 16.64. Although the interval is very wide, the entire range is
above the hypothesized value of zero. Therefore, at the 5% significance level the
sample data suggest that the mean difference differs from zero. In other words, there
has been a change in productivity due to the different layout in the workstation.
T E S T S T A T I S T I C F OR H Y P O T H E S I S T E S T S A B O U T μD
The test statistic for hypothesis tests about μD is assumed to follow the tdf distribu
tion with d f = n – 1, and its value is
where and sD are the mean and standard deviation, respectively, of the n sample
differences, and d0 is a given hypothesized mean difference.
EXAMPLE 10.7
Let’s revisit the chapter’s introductory case. Recall that a local ordinance requires
chain restaurants to post caloric information on their menus. A nutritionist wants to
examine whether average drink calories declined at Starbucks after the passage of
the ordinance. The nutritionist obtains transaction data for 40 Starbucks cardhold
ers and records their average drink calories prior to the ordinance and then after
the ordinance. A portion of the data is shown in Table 10.4. The entire data set can
be found on the text website, labeled Drink Calories. Using the critical value ap
proach, can she conclude at the 5% significance level that the ordinance reduced
average drink calories?
Similarly, in the fifth column of Table 10.4, we square the differences between di
and Summing these squared differences yields the numerator in the formula for
the sample variance s2D . The denominator is simply n – 1, so:
As usual, the standard deviation is the positive square root of the sample
variance, that is, We compute the value of the tdf test statistic
with d f = n – 1 = 40 – 1 = 39 as
H 0 : μD ≤ 0
H A : μD > 0
FIGURE 10.2 Excel's dialog box for t test with paired sam ple
S Y N O P S I S OF I N T R O D U C T O R Y CASE
In a n e ff o r t t o m a k e it e a s i e r for c o n s u m e r s t o s e l e c t h e a l t h i e r o p t i o n s ,
t h e g o v e r n m e n t w a n t s c h a i n r e s t a u r a n t s t o p o s t c a lo r ic i n f o r m a t io n o n
t h e ir m e n u s . A n u t r itio n is t s t u d i e s t h e e f f e c t s o f a r e c e n t local m e n u
o r d i n a n c e r eq u ir in g c a loric p o s t i n g s at a S t a r b u c k s in S an M a t e o , Cali
fornia. S h e o b t a i n s t r a n s a c t i o n d a t a for 4 0 S ta r b u c k s c a r d h o l d e r s a n d
r e c o r d s t h e ir a v e r a g e d rink a n d f o o d c a lo r i e s prior t o t h e o r d i n a n c e a n d
t h e n a fte r t h e o r d i n a n c e . T w o h y p o t h e s i s t e s t s are c o n d u c t e d . T h e first
t e s t e x a m i n e s w h e t h e r drink c a lo r ic in t a k e is l e s s s i n c e t h e p a s s a g e o f
t h e o r d i n a n c e . After c o n d u c t i n g a t e s t o n t h e m e a n d i f f e r e n c e a t t h e
5% s i g n i f i c a n c e leve l, t h e n u t r itio n ist infers t h a t t h e o r d i n a n c e d id n o t
p r o m p t c o n s u m e r s t o r e d u c e t h e ir drink c a lo r ic i n ta k e . T h e s e c o n d t e s t
i n v e s t i g a t e s w h e t h e r f o o d c a loric in t a k e is le s s s i n c e t h e p a s s a g e o f t h e o r d i n a n c e . At t h e
5% s i g n i f i c a n c e leve l, t h e s a m p l e d a t a s u g g e s t t h a t c o n s u m e r s h a v e r e d u c e d th e ir f o o d
c a lo r ic in t a k e s i n c e t h e p a s s a g e o f t h e o r d i n a n c e . In s u m , w h i l e t h e g o v e r n m e n t is tr y in g
t o e n s u r e t h a t c u s t o m e r s p r o c e s s t h e c a lo r ie i n f o r m a t io n a s t h e y are o r d e r in g , t h e r esults
a r e c o n s i s t e n t w i t h r e s e a r c h t h a t h a s s h o w n m i x e d r es u lts o n w h e t h e r m a n d a t o r y calo ric
p o s t i n g s are p r o m p t i n g c o n s u m e r s t o s e l e c t h e a l t h i e r f o o d s .
EXERCISES 10.2
S a m p l e results:
a. Specify the null and alternative hypotheses to test the
a. C a lc u la te t h e v a l u e o f t h e t e s t s t a t is t ic u n d e r t h e manager's assertion.
a s s u m p t i o n t h a t t h e d i f f e r e n c e is n o r m a ll y d is t r ib u t e d .
b. Assuming that completion times are normally
b. At t h e 5% s i g n i f i c a n c e le ve l, w h a t is t h e critical v a lu e ? distributed, calculate the value of the test statistic.
c. W h a t is t h e c o n c l u s i o n t o t h e h y p o t h e s i s t e s t? c. At the 10% significance level, what is the critical value(s)?
2 1 . C o n s id e r t h e f o l l o w i n g c o m p e t i n g h y p o t h e s e s a n d What is the decision rule?
a c c o m p a n y i n g s a m p l e result s. d. Is the manager's assertion supported by the data?
H y p o t h e s e s : H 0 : μD ≤ 0 ; H A : μD > 2 25. A diet center claims that it has the most effective weight loss
S a m p l e results: program in the region. Its advertisements say "Participants
in our program lose more than 5 pounds within a month."
a. C a lc u la te t h e v a l u e o f t h e t e s t sta tis tic . A s s u m e n o r m a li t y
Six clients of this program are weighed on the first day of the
t o a p p r o x i m a t e t h e p -v a lu e .
diet and then one month later.
b. U s e t h e 1 % s i g n i f i c a n c e le v e l to make a conclusion.
22. A sample of 35 paired observations generates the following W eight on First W eight One
resu lts: Client Day o f Diet M onth Later
a. Specify the appropriate hypotheses to test if the mean 1 158 151
difference is greater than zero. 2 205 200
b. Compute the value of the test statistic and approximate 3 170 169
the p -value.
4 189 179
c. At the 5% significance level, can you conclude that the
5 149 144
mean difference is greater than zero? Explain.
6 135 129
d. R e p e a t t h e h y p o t h e s i s t e s t u s in g t h e critical v a lu e ap p ro a c h .
23. Consider the following matched samples representing a. Specify the null and alternative hypotheses that test the
observations before and after an experiment. Assume diet center's claim.
that sample data are drawn from tw o normally distributed b. Assuming that weight loss is normally distributed,
populations. calculate the value of the test statistic.
a. Construct the competing hypotheses to determine 26. A bank employs tw o appraisers. When approving borrowers
if the experiment increases the m agnitude of the for mortgages, it is imperative that the appraisers value the
observations. sam e types of properties consistently. To make sure that
b. Implement the test at a 5% significance level. this is the case, the bank evaluates six properties that the
c. Do the results change if w e implement the test at a 1% appraisers had valued recently.
significance level?
Value from Value from
Property Appraiser 1 Appraiser 2
Applications 1 $235,000 $239,000
24. A manager of an industrial plant asserts that workers on 2 195,000 190,000
average do not complete a job using Method A in the same 3 264,000 271,000
amount of time as they would using Method B. Seven
4 315,000 310,000
workers are randomly selected. Each worker's completion
5 435,000 437,000
time (in minutes) is recorded by the use of Method A and
Method B. 6 515,000 525,000
a. Specify the com petin g hypotheses that determine a mock SAT, com plete a three-m onth test-prep course, and
whether there is any difference betw een the average then take the real SAT. These scores can also be found on the
values estimated by appraiser 1 and appraiser 2. text website, labeled M ock SAT.
b. Assuming that the difference is normally distributed, Student Score on Mock SAT Score on Real SAT
calculate the appropriate value of the test statistic. 1 1830 1840
c. Is there sufficient evid ence to conclude at the 5% 2 1760 1800
significance level that the appraisers are inconsistent in
3 2000 2010
their estimates? Explain.
4 2150 2190
27. FILE (Use Excel) Researchers at The Wharton School of
5 1630 1620
Business have found that m en and w o m e n shop for different
6 1840 1960
reasons (Knowledge@Wharton, November 2 8 , 2007). While
7 1930 1890
w o m e n enjoy the shoppin g experience, men are on a
mission to g e t the job done. Men do not shop as frequently 8 1710 1780
but w hen they do, they make big purchases like expensive
a. Specify the competing hypotheses that determine
electronics. Recently a large electronics store in Chicago
w hether completion of the test-prep course increases a
has started making eye-catching ads aimed to attract men's
student's score on the real SAT.
attention. In order to test the effectiveness of the strategy,
b. Assuming that SAT scores are normally distributed, calculate
the store records the number of daily male customers for
the value of the test statistic and its associated p -value.
10 days before and after the ads appeared. The data, sh own
c. At the 5% significance level, d o e s the sample data
in the accompanying table, can also be found on the text
support the test-prep providers' claims?
website, labeled M en a n d Electronics.
29. FILE (Use Excel) The following table shows the annual
Day Before After
returns (in percent) for Fidelity's Select Electronic and Select
1 38 42
Utilities mutual funds. These data can also be found on the
2 45 42
text website, labeled Electronic a n d Utilities.
3 49 46
Year Electronic Utilities
4 54 62
2001 − 14.23 − 21.89
5 62 70
2002 − 50.54 − 30.40
6 84 98
2003 71.89 26.42
7 77 88
2004 − 9.81 24.22
8 42 40
2005 15.75 9.36
9 44 42
2006 0.30 30.08
10 48 50
2007 4.67 18.13
a. Specify the com peting hypotheses to test the effectiveness 2008 − 49.87 − 36.00
of the strategy to attract more male customers. 2009 84.99 14.39
b. Compute the value of the test statistic, assuming that the
Source: w w w .finance.yahoo.com
number of male customers follows a normal distribution.
a. Set up the h ypotheses to test the claim that the mean
c. Interpret the results o f the test conducted at the 1% and
return for the Electronic mutual fund differs from the
the 10% levels of significance.
mean return for the Utilities mutual fund.
28. ES (Use Excel) A recent report criticizes SAT-test- b. Using the appropriate com m ands in Excel, find the value
preparation providers for promising big score gains without of the test statistic. What is the p -value?
any hard data to back up such claims (The Wall Street Journal,
c. At the 5% significance level, do the mean returns differ?
May 2 0 , 2009). Suppose eight college-bound students take
C O N F ID E N C E IN T E R V A L FOR p 1 − p 2
A 100(1 − α )% confidence interval for the difference between two population pro
portions p 1 − p 2 is given by:
As mentioned earlier, the above formula is valid for large samples; the general guideline
is that n1p1, n 1(1 − p 1), n2p 2, and n2{ 1 − p 2) must all be greater than or equal to 5, where
p 1 and p 2 are evaluated at respectively.
EXAMPLE 1 0.9
Despite his inexperience and some perception that he is a risky choice, candidate A
appears to have gained support among the electorate. Three months ago, in a survey
of 120 registered voters, 55 said that they would vote for Candidate A. Today,
41 registered voters in a sample of 80 said that they would vote for Candidate A.
Construct a 95% confidence interval for the difference between the two population
proportions.
SOLUTION: Let p 1 and p 2 represent the population proportion of the electorate
who support the candidate today and three months ago, respectively. In order to
calculate a 95% confidence interval for p 1 − p 2, we use the formula
We compute the sample proportions as
With 95% confidence, we can report that the change in the proportion of support for
the candidate is between − 8.70% and 19.54%.
For notational simplicity, we again use the symbol d0 to denote a given hypothesized dif
ference, but now d0 is not a hypothesized mean difference, but a hypothesized difference
between the unknown population proportions P1 and p2. In most cases, d0 is set to zero.
For example, when testing if the population proportions differ— that is, if p 1 ≠ p 2— we
use a two-tailed test with the null hypothesis defined as H0: p 1 − p 2 = 0 . If, on the other
hand, we wish to determine whether or not the proportions differ by some amount, say
20%, we set d0 = 0.20 and define the null hypothesis as H0: p 1 − p 2 = 0.20. One-tailed
tests are defined similarly.
EXAMPLE 1 0 .1 0
Let’s revisit Example 10.9. Specify the competing hypotheses in order to determine
whether the proportion of those who favor Candidate A has changed over the three-
month period. Using the 95% confidence interval, what is the conclusion to the test?
Explain.
where
2. If the hypothesized difference d0 is not zero, then the value of the test statistic is
EXAMPLE 10.1 1
Recent research by analysts and retailers claims significant gender differences
when it comes to online shopping (The Wall Street Journal, M arch 13, 2008).
A survey revealed that 5,400 o f 6,000 men said they “regularly” or “occasion
ally” make purchases online, com pared with 8,600 o f 10,000 women surveyed.
At the 5% significance level, test whether the proportion of all men who regu
larly or occasionally make purchases online is greater than the proportion of
all women.
SOLUTION: We use the critical value approach to solve this example. Let p 1 and
p 2 denote the population proportions of men and of women who make online
purchases, respectively. We wish to test whether the proportion of men who make
purchases online is greater than the proportion of women, or p 1 − p 2 > 0. Therefore,
we construct the competing hypotheses as
H0: p1 − p2 ≤ 0
HA: p1 − p2 > 0
For a right-tailed test with α = 0.05, the appropriate critical value is z α = Z0.05 =
1.645. The decision rule is to reject H0 if z > 1.645. Since 7.41 > 1.645, we reject
H(). The proportion of men who shop online either regularly or occasionally is
greater than the proportion of women at the 5% significance level. Our results are
consistent with the recent decision by so many retailers to redesign their websites to
attract and keep male customers.
EXAMPLE 10.12
While we expect relatively expensive wines to have more desirable characteris
tics than relatively inexpensive wines, people are often confused in their assess
ment of the quality of wine in a blind test (New York Times, December 16, 2010).
In a recent experiment at a local winery, the same wine is served to two groups
of people but with different price information. In the first group, 60 people are
told that they are tasting a $25 wine, of which 48 like the wine. In the second
group, only 20 of 50 people like the wine when they are told that it is a $10 wine.
The experiment is conducted to determine if the proportion of people who like
the wine in the first group is more than 20 percentage points higher than those in
the second group. Conduct this test at the 5% significance level using the p -value
approach.
SOLUTION: Let p 1 and p 2 denote the proportions of people who like the wine in
groups 1 and 2, respectively. We want to test if the proportion of people who like the
wine in the first group is more than 20 percentage points higher than in the second
group. Thus, we construct the competing hypotheses as
H0: p 1 − p 2 ≤ 0.20
HA: p1 − p2 > 0.20
For this right-tailed test, we compute the p -value as P (Z ≥ 2.31) = 1 − 0.9896 = 0.0104.
Since the p -value < α , or 0.0104 < 0.05, we reject the null hypothesis. At the 5%
significance level, we conclude that the proportion of people who like the wine in
the first group is more than 20 percentage points higher than in the second group.
Overall, this result is consistent with scientific research, which has demonstrated the
power of suggestion and expectations in wine tasting.
EXERCISES 10.3
31. G iv e n c onstruct
a 9 0 % c o n f i d e n c e in t er v a l fo r t h e d i f f e r e n c e b e t w e e n
Applications
t h e p o p u l a t i o n p r o p o r t i o n s . Is t h e r e a sta tis tic a l d i f f e r e n c e 36. A recent study claims that girls and boys do not do equally
b e t w e e n t h e p o p u l a t i o n p r o p o r t i o n s a t t h e 10% s i g n i f i c a n c e well on math tests taken from the 2nd to 11th grades (The
le v e l? Explain. Chicago Tribune, July 2 5 , 2008). Suppose in a representative
sample, 344 of 43 0 girls and 369 of 450 boys score at
32. Consider the following com peting hypothese s and
proficient or advanced levels on a standardized math test.
accompanying sample data.
a. Construct a 95% confidence interval for the difference
H0: P1 − P2 = 0
betw een the population proportions of boys and girls
HA: P1 − P2 ≠ 0 w h o score at proficient or advanced levels.
H0: P1 − P2 ≥ 0 o u t s i d e t h e ir r a ce o r e t h n ic i ty . S u p p o s e b o t h o f t h e s e s u r v e y s
w e r e c o n d u c t e d o n 1 2 0 n e w l y m a r r ie d c o u p l e s .
HA: P1 − P2 < 0
a. Specify the competing hypotheses to test the claim that
x1 = 250 x2 = 275
there is an increase in the proportion of people w h o
n1 = 400 n2 = 400 marry outside their race or ethnicity.
a. Calculate the value of the test statistic. b. What is the value of the test statistic and the associated
b. At t h e 5 % s i g n i f i c a n c e le vel, c a l c u l a t e t h e critical v a lu e (s ). p -value?
c. W h a t is t h e c o n c l u s i o n ? c. At the 5% level of significance, what is the conclusion?
34. Consider the following competing hypotheses and 38. Research by Harvard Medical School experts su ggests that
accompanying sample data. boys are more likely than girls to grow out of childhood
H0: P1 − P2 = 0 asthma w hen they hit their te en a g e years (BBC News, August
1 5 , 2008). Scientists followed over 1,000 children between
HA: P1 − P2 ≠ 0
the ages of 5 and 12, all of w h om had mild to moderate
x1 = 300 x2 = 325 asthma. By the a g e of 1 8 , 14% of the girls and 27% of the
As college costs soar, costs of living rise, and student aid continues to drop,
more and more college students rely on credit cards to pay for college ex
penses. A recent study reports a massive increase in credit card debt among
college students over the last four years (Money, April 15, 2009). Timothy
McKenna, dean of student affairs at a local university, has been studying
this trend with his own undergraduates. In 2007, he surveyed 50 graduat
ing seniors and found that their average credit card debt was $2,877 with a
standard deviation of $375. The first column of Table 10.6 shows a portion
of the data; the entire data set labeled Student Debt can be found on the text
website. He also found that 28 out of 50 freshmen had substantial financial aid. Repeating
the survey in 2011, Timothy surveys 50 graduating seniors and finds that these students
have average credit card debt of $3,926 with a standard deviation of $471. A portion of
these data is shown in the second column of Table 10.6. In addition, 12 out of 50 freshmen
had substantial financial aid.
TABLE 1 0 . 6 Senior Credit Card Debt, 2007 versus 2011
$2,939 $4,422
2,812 4,266
⋮ ⋮
2,856 3,718
Timothy wants to use the sample data to:
1. Determine whether average credit card debt was significantly lower for seniors in
2007 as compared to 2011.
2. Determine whether the proportion of freshmen with substantial financial aid was sig
nificantly greater in 2007 as compared to 2011.
Rising education costs, mounting loan debt, and a slumping student job market have Sample Report
combined to create a dim financial picture for undergraduates after graduation. More
undergraduates are using their credit cards to pay for college expenses such as texts and — Credit Card
fees (USA Today, April 13, 2009). This usage is resulting in higher and higher unpaid
balances. A university dean wonders whether this trend is mirrored at his university. In Debt of
2007, he surveyed 50 graduating seniors and found that their average credit card debt was
$2,877 with a standard deviation of $375. In 2 0 1 1, he took another sample of 50 gradu Undergraduates
ating seniors and found that their average credit card debt was $3,926 with a standard
deviation of $471. A hypothesis test was conducted at the 5% significance level in order
to determine whether average credit card debt in 2007, denoted μ 1 , was significantly less
than average credit card debt today, denoted μ 2. Given that undergraduate spending pat
terns may have changed over the last few years, the population variances are not assumed
equal. The second row of Table 10. A shows the appropriate competing hypotheses, the
value of the test statistic, and the p -value for this test. Since the p -value is approximately
0.0000, there is enough evidence to reject the null hypothesis and conclude that average
credit card debt for graduating seniors in 2007 was less than it is in 2011.
TABLE 1 0 .A T e s t S t a t is t ic s a n d p - V a lu e s fo r H y p o t h e s i s T e s ts
H0: μ1 − μ2 ≥ 0
HA: μ1 − μ2 < 0 0.0000
H0: P1 − P2 ≤ 0
HA: P1 − P2 > 0 0.0005
The dean then tested whether the proportion of freshmen with substantial financial
aid in 2007, denoted p 1, is greater than the proportion of freshmen with substantial
financial aid today, denoted p 2. The dean’s survey found that 28 out of 50 freshmen
had substantial financial aid in 2007, yielding a sample proportion
in 2011, the survey showed that the sample proportion of freshmen with substantial
financial aid was The third row of Table 10. A shows the compet
ing hypotheses, the value of the test statistic, and the p -value for this test. At the 5%
significance level, the proportion of freshmen with substantial financial aid in 2007
was greater than the proportion in 2011. The results of both tests are consistent with
the findings nationwide. In sum, there has been a substantial rise in credit card debt
of undergraduates, much of it fueled by the rising cost of college and a financial aid
system that cannot keep up.
Conceptual Review
LO 10.1 M ake inferences about the difference betw een tw o population means based on
independent sampling.
Independent samples are samples that are completely unrelated to one another.
A 100(1 − α )% confidence interval for the difference between two population
means μ1 − μ 2 ,, based on independent samples, is
are known.
The degrees of freedom for the last two tests are the same as the ones defined for the
corresponding confidence intervals.
LO 1 0.2 M ake inferences about the mean difference based on matched-pairs sampling.
In general, two types of matched-pairs sampling occur. The first type of a matched-
pairs sample is characterized by a measurement, an intervention of some type, and then
another measurement. We generally refer to these experiments as “before” and “after”
studies. The second type of a matched-pairs sample is characterized by a pairing of
observations, where it is not the same individual that gets sampled twice.
For matched-pairs sampling, the population parameter of interest is referred to as the
mean difference μD where D = X 1 − X 2, and the random variables X 1 and X2 are
matched in a pair. A 100(1 − α )% confidence interval for the mean difference μD,
based on a matched-pairs sample, is given by and sD are the
mean and the standard deviation, respectively, of D , and d f = n − 1. When conducting
a hypothesis test about μD the value of the test statistic is calculated as
where d0 is a hypothesized mean difference.
LO 1 0.3 M ake inferences a bout the difference between tw o population proportions based
on independent sampling.
A 100(1 − α )% confidence interval for the difference between two population
proportions p 1 − p 2 is given by
When conducting hypothesis tests about the difference between two proportions
P 1 − p 2, the value of the test statistic is calculated as:
if the hypothesized difference d0 between p 1 and p 2 is zero.
not zero.
Additional Exercises and Case Studies
42. Do men really spen d more m oney on St. Patrick's Day as b. Assuming that crop yields are normally distributed,
compared to wom en? A recent survey found that men spend calculate the value of the test statistic.
an average of $43.87 while w om e n spend an average of c. At the 5% significance level, find the critical value(s).
$29.54 (USA Today, March 1 7 , 2009). Assume that these data d. Is there sufficient evident to conclude that the crop
were based on a sample of 100 men and 100 w o m e n and yields are different? Should the farmer be concerned?
the population standard deviations of spending for men and
45. FIL E (Use Excel) The accompanying table shows annual
w om en are $32 and $25, respectively.
return data from 2001– 2009 for Vanguard's Balanced Index
a. Specify the com peting hypotheses to determine
and European Stock Index mutual funds. The data can
whether men spend more m oney than w om en spend on
also be accessed from the text website, labeled Vanguard
St. Patrick's Day.
Balanced a n d European Funds.
b. Calculate the value o f the test statistic.
c. Calculate the p -value. European Stock
d. At the 1% significance level, do men spend more money Year Balanced Index Index
than w om e n spend on St. Patrick's day? Explain. 2001 − 3.02% − 20.30%
43. A new study has found that, on average, 6- to 12-year-old 2002 − 9.52 − 17.95
children are spending less time on household chores today 2003 19.87 38.70
compared to 1981 levels (The Wall Street Journal, August 27, 2004 9.33 20.86
2008). Suppose two samples representative of the study's 2005 4.65 9.26
results report the following summary statistics for the two
2006 11.02 33.42
periods:
2007 6.16 13.82
1981 Levels 2008 Levels 2008 − 22.21 − 44.73
2009 20.05 31.91
s1 = 4.2 m in u te s s2 = 3.9 minutes
Source: w w w .finance.yahoo.com
n1 = 30 n2 = 30
a. Set up th e hypotheses to test w h e th e r m ean returns o f
a. S p e c i f y t h e c o m p e t i n g h y p o t h e s e s t o t e s t t h e s tu d y ' s
th e tw o fu nds differ.
c la im t h a t c h ild r e n t o d a y s p e n d le s s t i m e o n h o u s e h o l d
b. W h a t is th e value o f th e te st statistic and its associated
c h o r e s a s c o m p a r e d t o c h ild r e n in 1 9 8 1 .
p -value given unequal p o p u la tio n standard deviations?
b. C a lc u la t e t h e v a l u e o f t h e t e s t s ta tis tic a s s u m i n g t h a t t h e
c. At th e 5% significance level, w h a t is th e conclusion?
u n k n o w n p o p u la tio n variances are e qual.
46. A recent Health o f Boston re p o rt suggests th a t 14% o f fem ale
c. Approximate the p -value.
residents suffer from asthma as opposed to 6% o f males (The
d. At the 5% significance level, do the data support the
Boston Globe, August 16, 2010). Suppose 250 females and
study's claim? Explain.
200 males responded to the study.
e. Repeat the hypothesis test using the critical value
a. Develop the appropriate null and alternative hypotheses
approach.
to test whether the proportion of females suffering from
4 4 . A f a r m e r is c o n c e r n e d t h a t a c h a n g e in f erti li zer t o a n o r g a n i c asthma is greater than the proportion of males.
v a r ia n t m i g h t c h a n g e his c r o p y ie ld . He s u b d i v i d e s 6 lo ts a n d
b. Calculate the value of the test statistic and its associated
u s e s t h e o ld fe rtiliz er o n o n e h a lf o f e a c h lot a n d t h e n e w
p -value.
ferti li zer o n t h e o t h e r half. T h e f o l l o w i n g t a b l e s h o w s t h e
c. At the 5% significance level, what is the conclusion?
res ults .
Do the data suggest that females suffer more from
Crop Yield Using Crop Yield Using asthma than males?
Lot Old Fertilizer New Fertilizer 47. Depression engulfs millions of Americans every day. A new
1 10 12 federal study reported tha t 10.9% o f adults aged 18–24
2 11 10 identified w ith some level of depression versus 6.8% of
finds that 70% of t h e flights were on tim e at JFK compared 50. FILE (Use Excel) An engineer wan ts to determine the
with 63% at O'Hare. Su pp ose t h e s e proportions were effectiveness of a safety program. He collects annual loss of
based on 200 flights at each of t h e s e t w o airports. The hours due to accidents in 12 plants "before and after" the
analyst be lieves that th e proportion o f on-tim e flights at program was put into operation. The data can be found on
JFK is more than 5 percentage points higher than that of the text website, labeled S afety Program .
O'Hare. Plant Before After Plant Before After
a. Develop the com petin g hypotheses to test the 1 100 98 7 88 90
transportation analyst's belief.
2 90 88 8 75 70
b. Compute the value of the test statistic.
3 94 90 9 65 62
c. Use th e p -value approach to test the above assertion.
4 85 86 10 58 60
d. Repeat the analysis with the critical value approach.
5 70 67 11 67 60
49. FILE (Use Excel) The SAT is required of m ost students 6 83 80 12 104 98
applying for college admission in the United States. This
standardized test has g o n e through many revisions over a. Specify the competing hypothese s that determine
the years. In 2005, a n e w writing section was introduced whether the safety program w as effective.
that includes a direct writing measure in the form of an b. Assuming that hours are normally distributed, calculate
essay. People argue that female stu dents generally do the value of the test statistic.
worse on math tests but better on writing tests. Therefore, c. At th e 5% significance level, calculate the critical
the n e w section may help reduce th e usual male lead on value(s).
the overall average SAT score (The Washington Post, August d. Is there sufficient evident to conclude that the safety
3 0 , 2006). Consider the following scores on th e writing program was effective? Explain.
CASE STUDIES
D a ta f o r C a s e S tu d y 1 0 .1 A n n u a l R e tu r n s for
F ir m s in H e a l t h a n d I n f o r m a t i o n T e c h n o l o g y I n d u s t r ie s
10.29% 4.77%
32.17 1.14
⋮ ⋮
13.21 22.61
In a report, use the sample information to:
1. Provide descriptive statistics and comment on the reward and risk in each industry.
2. Determine whether the average returns in each industry differ at the 5% significance
level. Assume that the population variances are unequal.
LEARNING OBJECTIVES
T E R
Statistical Inference
C oncerning Variance
A fter rea d in g LO 1 1 .1 D iscu ss fe a tu r e s o f t h e χ2 d istrib u tio n .
th is ch a p ter
y o u sh o u ld LO 1 1 .2 C onstruct a c o n fid e n c e in terval for th e p o p u la tio n varian ce.
b e a b le t o :
LO 1 1 .3 C on d u ct a h y p o th e s is t e s t for t h e p o p u la tio n varian ce.
A s y n o p s i s o f th is c a s e is p r o v i d e d at t h e e n d o f S e c t i o n 11.2.
11.1 Inference Concerning the Population
Variance
The population variance is used in quality-control studies to measure the variability of
the weight, size, or volume of the product. Consider, for example, a bottler who wishes
its production line to fill a certain amount of beverage in each bottle. It is important
not only to get the desired average amount filled in the bottles, but also to keep the
variability of the amount filled below some tolerance limit. Similarly, with variance
used as a quantitative measure of risk, an investor may want to evaluate his/her risk in
a particular investment. Other examples for the relevance of making inference regard
ing the population variance include evaluating the consistency of an athlete or a team,
the variability of speeds on a highway, and the variability of repair costs of a certain
automobile.
Recall that we use the sample mean as the estimator of the population mean μ .
Similarly, we use the sample variance S2 as an estimator of the population variance σ 2.
Like S2 has the desirable properties of an estimator. In particular, it is unbiased
(E(S2) = σ 2), efficient, and consistent. Using a random sample of n observations drawn
from the population, we compute as an estimate of σ 2. In order to examine
the techniques for statistical inferences regarding σ 2, we first need to analyze the sam
pling distribution of S2.
T H E S A M P L I N G D I S T R I B U T I O N OF
If a sample of size n is taken from a normal population with a finite variance, then
the statistic follows the χ2df distribution with d f = n − 1.
In earlier chapters, we denoted the random variables by upper-case letters and par
ticular outcomes o f the random variables by the corresponding lower-case letters. For
instance, the statistics Z and Tdf are random variables and their values are given by
z and tdf, respectively. It is cumbersome to continue with the distinction between the
random variable and its value in this chapter. Here we use the above statistic χ2df to
represent a random variable as well as its value. Similarly, for the distribution
introduced in Section 11.2, we will use to represent both a random variable
and its value.
From Figure 11.1, we note that the χ2df distributions are positively skewed, where the
extent of skewness depends on the degrees of freedom. As the d f grow larger, the χ2df dis
tribution tends to the normal distribution. For instance, in the figure, the χ220 distribution
resembles the shape of the normal distribution.
FIGURE 11.1 The χ2df distribution with various d egr ee s of
freedom
S U M M A R Y OF T H E χ2d
f DIS TR IB U TION
A portion of the right-tail probabilities and the corresponding values of χ2df distributions
are given in Table 11.2. Table 3 of Appendix A provides a more complete table.
Suppose we want to find the value χ2α,df with α = 0.05 and d f = 10, that is, χ20.05,10. Using
Table 11.2, we look at the first column labeled d f and find the value 10. We then continue
along this row until we reach the column χ20.050. Here we see the value χ20.05,10 = 18.307
such that P(χ210 > 18.307) = 0.05.
TABLE 11.2 Portion of the χ2
S ta b le
Area in Upper Tail, α
df 0.995 0.990 0.975 0.950 0.900 0.100 0.050 0.025 0.010 0.005
1 0.000 0.000 0.001 0.004 0.016 2.706 3.841 5.024 6.635 7.879
⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮
10 2.156 2.558 3.247 3.940 4.865 15.987 18.307 20.483 23.209 25.188
⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮
100 67.328 70.065 74.222 77.929 82.358 118.342 124.342 129.561 135.807 140.170
Sometimes we need to derive values in the left (lower) tail of the distribution. Given that
the area under any probability distribution equals one, if the area to the left of a given
value equals α , then the area to the right must equal 1 − α . In other words, the relevant
value on the left tail of the distribution is ,dfwhere
1−
χ2
α P(χ2df ≥ χ21 − α,df) = 1 - α .
L O C A T I N G χ2df- V A L U E S O N T H E L E F T T A I L
For a χ2df distributed random variable, χ21 − α,df represents a value such that
P(χ2df < χ21 − α,df) = α, or equivalently, P(χ2df ≥ χ21 − α,df) = 1 − α.
Suppose that we want to find the value such that the area to the left of the χ210 variable
equals 0.05. Given that the area to the left of this value is 0.05, we know that the area to
its right is 1 − 0.05 = 0.95; thus, we need to find χ21−0 .05,10 = χ20.95,10. Again we find d f = 10
in the first column and follow this row until we intersect the column χ20.95 and find the
value 3.940. This is the value such that P(χ210 ≥ 3.940) = 0.95 or P(χ210 < 3.940) = 0.05.
Figure 11.3 graphically depicts the probability α = 0.05 on both sides of the χ210 distribu
tion and the corresponding χ210-values.
EXAMPLE 11.1
Find the value x for which:
SOLUTION:
a. We find the value x such that the probability in the right tail of the distribution
equals 0.025. Therefore, x = 12.833.
b. We find the value x such that the probability in the left tail of the distribution
equals 0.025. We solve this problem as P(χ28 ≥ x) = 1 − 0.025 = 0.975 where
x = 2.180. This is equivalent to P(χ28 < 2.180) = 0.025.
Confidence Interval for the Population Variance LO 11.2
Consider a χ2df distributed random variable. Using the notation introduced above, we can Construct a
make the following probability statement concerning this random variable: confidence interval
for th e population
P (χ2 1 − α/2 ,d f ≤ χ2 d f ≤ χ2 α/2 ,d f ) = 1 − α. variance.
This indicates that the probability that χ2df falls between χ21 − α/2,df and χ2α/2,df is equal to
1 − α , where 1 − α is the familiar confidence coefficient. For instance, for a 95%
confidence level, a = 0.05, χ2α/2,df = χ20.025,df and χ21 − α/2,df = χ20.975,df. Substituting
into the probability statement yields
After manipulating this equation algebraically, we arrive at the formula for the confidence
interval for σ 2 .
C O N F ID E N C E IN TER VA L FOR σ2
A 100(1 − α )% confidence interval of the population variance σ2 is computed as
where d f = n − 1.
This formula is valid only when the random sample is drawn from a normally distributed
population. Note that the confidence interval is not in the usual format of point estimate ±
margin of error. Since the confidence intervals of the population mean and the population
proportion are based on the z or the tdf distributions, the symmetry of these distributions
leads to the same margin of error that is added to and subtracted from the point estimate.
However, for a nonsymmetric χ2df distribution, what is added to the point estimate of the
population variance is not the same as what is subtracted. Finally, since the standard de
viation is just the positive square root of the variance, a 100(1 − α )% confidence interval
for the population standard deviation is computed as
EXAMPLE 11.2
Compute 95% confidence intervals for the population standard deviation for the
Metals fund as well as the Income fund using the data from Table 11.1 in the intro
ductory case. Assume that returns are normally distributed.
Here σ20 is the hypothesized value of the population variance σ 2 . As before, we can use
the confidence intervals to implement two-tailed hypothesis tests; however, we apply
the four-step procedure to conduct one-tailed hypothesis tests concerning the population
variance.
We use the sample variance to conduct hypothesis tests regarding the population vari
ance. As noted earlier, we use a χ 2test if the underlying population is normally distributed.
EXAMPLE 11.3
We again consider the introductory case. Rebecca Johnson’s client asks if the stan
dard deviation of returns for the Metals fund is significantly greater than 25%. This
is equivalent to testing whether or not the variance is significantly greater than
625(%)2. We conduct this test at the 5% significance level, based on the sample
information provided in Table 11.1. We implement the four-step procedure using
the critical value approach and assume that returns are normally distributed.
EXAMPLE 11.4
Rebecca Johnson’s client from the introductory case also wonders if the standard
deviation of returns for the Income fund differs from 8.5%. This is equivalent to
testing whether or not the variance differs from 72.25(%)2. Use the p-value approach
to conduct this test at the 5% significance level.
Recall that the p-value is the probability of obtaining a value of the test statistic
that is at least as extreme as the one that we actually observed, given that the null
hypothesis is true as an equality. For a two-tailed test, we double the probability that
is considered extreme. For example, for a two-tailed test of the population mean
μ with a z statistic, the p-value is computed as 2P(Z ≥ z) if z > 0 or 2P(Z ≤ z) if
z < 0. Note that z > 0 if and z < 0 if Similarly, for a two-tailed test
of the population variance σ 2, the p -value is computed as two times the right-tail
probability if s2 > σ2 0 or two times the left-tail probability if s2 > σ2 0.
For the Income fund, since s2 > σ2 0 (122.54 > 72.25), we compute the p -value
as 2P(χ2df ≥ 15.27). For d f = 9, since 15.27 lies between 14.684 and 16.919 (see
χ2 table), P(χ29 − 15.27) lies between 0.05 and 0.10. Multiplying this probability by
two results in a p -value between 0.10 and 0.20.
At the 5% significance level, we cannot reject H0 because the p -value is greater
than α = 0.05. Therefore, we cannot conclude that the risk, measured by the vari
ance of the return, differs from 72.25(%)2; or equivalently, we cannot conclude that
the standard deviation differs from 8.5%.
Excel returns the probability of 0.0838. The exact p -value for the two-tailed test is
2(0.0838) or 0.1676. Again, given a significance level of 5%, we are unable to reject the
null hypothesis that the population variance equals 72.25(%)2.
EXERCISES 11.1
deviation are calculated as 52.5 and 3.8, respectively. Assume
Concepts
that the population is normally distributed.
1. Find the value x for which:
a. Construct a 90% interval estim ate of the population
variance.
b. Construct a 99% interval estimate o f the population
variance.
c. Use your answers to discuss the impact of the confidence
2. Find the value x for which: level on the width of the interval.
5. The following values are drawn from a normal population.
20 29 32 27 34 25 30 31
H A : σ2 ≠ 200 Year E n e rg y
Approxim ate the p -value based on the follow ing sample 2005 44.60
inform ation, where the sample is drawn from a norm ally 2006 19 .6 8
distributed population. 2007 3 7 .0 0
a. s2 = 300; n = 25 2008 − 4 2 .8 7
b. s2 = 100; n = 25 2009 3 8 .3 6
c. Which o f the above sample inform ation enables us to Source: w w w .finance.yahoo.com .
reject the null hypothesis at α = 0.05?
a. C a lc u la te t h e p o i n t e s t i m a t e o f σ .
8. You w o u ld like to te st th e claim th a t th e variance o f a b. C o n s t r u c t a 9 5 % c o n f i d e n c e in ter v al o f σ .
n o rm a lly d is trib u te d p o p u la tio n is m ore th a n 2 squared
1 3 . T h e m a n a g e r o f a s u p e r m a r k e t w o u l d lik e t h e v a r i a n c e
units. You d ra w a random sam ple o f 10 o b se rva tio n s as 2 , 4,
o f th e w a itin g tim e s o f th e c u sto m e r s n o t t o e x c e e d
1 , 3 , 2 , 5 , 2 , 6 , 1 , 4. A t α = 0.10, te s t th e above claim using
3 m in u te s-sq u a r e d . S h e w o u ld a d d a n e w ca sh register
(a) th e p -value approach and (b) th e c ritic a l value approach.
if t h e v a r i a n c e e x c e e d s t h i s t h r e s h o l d . S h e r e g u l a r l y
ch eck s t h e w a itin g tim e s o f th e c u sto m e r s to en su r e th at
Applications
it s v a r i a n c e d o e s n o t r ise a b o v e t h e a l l o w e d l e v e l . In a
9. A research a nalyst is e xa m in in g a stock fo r possible in clu sio n
recen t ra ndom sa m p le o f 2 8 c u s to m e r w a itin g tim es,
in his client's p o rtfo lio . O ver a 10-year pe rio d , th e sam ple
s h e c o m p u t e s th e s a m p le v a rian ce as 4.2 m in u te s -
m ean and th e sam ple standard d e v ia tio n o f an n ua l returns
s q u a r e d . S h e b e lie v e s th a t t h e w a itin g tim e s are n o rm a lly
on th e stock w ere 20% and 15%, respectively. The c lie n t
d istrib u ted .
w a n ts to kn o w if th e risk, as m easured by th e standard
a. S t a t e t h e nu ll a n d t h e a lt e r n a t i v e h y p o t h e s e s t o t e s t if
d e v ia tio n , d iffe rs fro m 18%.
th e th reshold has b e e n crossed.
a. C o n stru ct a 95% co n fid en ce in te rv a l o f th e p o p u la tio n
b. U s e t h e p - v a l u e a p p r o a c h t o c o n d u c t t h e t e s t at
variance and th e n in fe r th e 95% co n fid en ce in te rv a l o f
α = 0.05.
th e p o p u la tio n standard d e v ia tio n .
c. R e p e a t t h e a n a ly s is w i t h t h e critical v a l u e a p p r o a c h .
b. W hat a ssu m p tio n d id you m ake in co n stru ctin g th e
d. W hat sh ou ld th e m a n a g er do?
co n fid en ce interval?
14. A r e s t a u r a n t o w n e r is c o n c e r n e d a b o u t t h e c o n s i s t e n c y o f
c. Based on th e above co n fid en ce in te rva l, does th e risk
b u s i n e s s . H e w a n t s t o d e t e r m i n e if t h e s t a n d a r d d e v i a t i o n o f
d iffe r fro m 18%?
t h e pr o fits fo r e a c h w e e k is l e s s t h a n $ 3 0 0 . T h e pr o fits fr o m
10. A rep la ce m e nt p a rt fo r a m achine m u st be p ro d u ce d w ith in
la st w e e k a r e li s te d b e l o w (in d o lla rs). A s s u m e t h a t pr o fits are
close sp e cifica tion s in o rd e r fo r it to be acceptable to
n o r m a ll y d is t r ib u t e d .
custom ers. Suppose th e sam ple variance fo r 20 parts tu rn s
1,825 1,642 1,675 1, 553 1,925 2 ,0 3 7 1,9 02
o u t to b e s2 = 0.03.
a. C o n stru ct a 95% co n fid en ce in te rv a l o f th e p o p u la tio n a. S t a t e t h e a p p r o p r i a t e nu ll a n d a lt e r n a t i v e h y p o t h e s e s for
variance. t h e test.
b. P ro d u ction sp e cifica tion s call fo r a variance in th e b. C o m p u t e t h e v a l u e o f t h e t e s t sta tis tic .
le n g th s o f th e parts to be e xa ctly 0.05. C om m ent on c. U s e t h e critical v a l u e a p p r o a c h t o t e s t t h e o w n e r 's
w h e th e r o r n o t th e sp e cifica tio n is b e in g vio la te d . co n c er n at a = 0.01.
c. T h e g r o u p d e c i d e s t o b e g i n a l o b b y i n g e f f o r t o n its
a. S ta te th e c o m p e tin g h y p o th e se s.
m e m b e r s ' b e h a l f if t h e v a r i a n c e in t h e pr ic e d o e s n o t
b. C o m p u t e t h e v a lu e o f t h e t e s t sta tistic. W h a t
e q u a l 4. W h a t s h o u l d t h e g r o u p d o ?
a ssu m p tio n r e g a r d i n g t h e IFN r e t u r n s d i d y o u m a k e?
c. G iv e n α = 0 . 0 5 , s p e c i f y t h e critical v a lu e (s ). 18 . F IL E (Use Excel) W hile the housing m arket has not recovered,
d. Is t h e v a r ia n c e o f r e t u r n s g r e a t e r th a n 1, 0 0 0 (% )2? real estate investm ent in college towns continues to promise
16. F IL E (U s e Excel) A realto r in M is s io n Viejo, California, b e l i e v e s good returns (The Wall Street Journal, September 24,
u n its , w h e r e e a c h u n it e q u a l s $ 1 , 0 0 0 . A s s u m e h o u s e p r ic e s but the same cannot be said fo r students. There also tends
are n o r m a ll y d is tr ib u te d .
to be significant variability in rents. Consider m onthly rents
o f tw o bedroom apartm ents in tw o campus towns: Ann
a. S t a t e t h e null a n d t h e a lt e r n a t i v e h y p o t h e s e s fo r t h e t e s t .
Arbor, M ichigan, and Davis, California. A po rtion o f the data
b. O p e n t h e M is s io n V iejo (r a w ) d a t a f r o m t h e t e x t w e b s i t e
is shown in the accom panying table; the com plete data,
in t o a n Excel s p r e a d s h e e t (d a ta a r e in $ 1 , 0 0 0 s ) . C a lc u la te
labeled Rentals, can be found on the te xt website.
t h e v a l u e o f t h e t e s t s ta tis tic .
t o c a l c u l a t e t h e p -v a lu e . $850 $744
d. At α = 0 . 0 5 w h a t is t h e c o n c l u s i o n ? Is t h e realtor's c la im 929 850
s u p p o r te d by th e data? ⋮ ⋮
1 7 . F I L E ( U s e Excel) A c c e s s t h e m i l e s p e r g a l l o n ( m p g ) d a t a o n 1450 1810
th e te x t w e b s ite lab eled MPG. Source: w w w .zillo w .co m .
The F distribution is nam ed in honor of Sir Ronald Fisher, who discovered the distribution in 1922.
positively skewed with values ranging from zero to infinity, but becomes increasingly
symmetric as d f 1and df2 increase.
In general, the distribution is the probability distribution of the ratio of two
independent chi-square variables, where each variable is divided by its own degrees of
freedom, that is,
T H E S A M P L I N G D I S T R I B U T I O N OF S21/S22 W H E N σ2 1 /σ2 2
If independent samples of size n1 and n2 are drawn from normal populations with
equal variances, then the statistic follows the distribution with
df1 = n 1 − 1 and df2 = n2 − 1.
As mentioned earlier, we will use to represent both a random variable and its value.
S U M M A R Y OF T H E DISTRIBUTION
From Figure 11.5 we note that all distributions are positively skewed, where skew
ness depends on degrees of freedom, df1 and df2 As df1 and df2 grow larger, the
distribution becomes less skewed and tends to the normal distribution. For instance, F(20,20)
is relatively less skewed and more bell-shaped as compared to F(2,8) or F(6,8).
A portion of the right-tail probabilities α and the corresponding values are given in
Table 11.3. Table 4 of Appendix A provides a more complete table.
T a b le 1 1 .3 P ortion o f t h e F Table
N u m era to r D e g r e e s o f F reed o m , d f 1
D e n o m in a to r D e g r e e s Area in U p p er
o f F reed o m , d f2 Tail, α 6 7 8
0.10 3.05 3.01 2.98
0.05 4.28 4.21 4.15
6
0.025 5.82 5.70 5.60
0.01 8.47 8.26 8.10
EXAMPLE 11.5
Find the value x for which:
SOLUTION:
a. We find the value x such that the probability under the right tail of the
distribution equals 0.025. Therefore, given P(F(7 10) ≥ x ) = 0.025, we
find x = 3.95.
b. We find the value x such that the probability under the left tail of the
distribution equals 0.05, or equivalently, the probability in the right tail of
the distribution equals 0.95. We have In other
words, P(F(7,10) < 0.27) = 0.05.
LO 11.5
Confidence Interval for the Ratio
Construct a
of Two Population Variances
confidence interval
The formula for a confidence interval for the ratio of the population variances σ21/ σ22 is for the ratio of
derived in a manner analogous to previous confidence intervals. Here, we simply show tw o population
the end result. variances.
C O N F ID E N C E INTERVAL FOR σ21/σ22
A 100( 1 − α )% confidence interval for the ratio of the population variances σ21/σ22
is computed as
This specification is based on the assumption that the sample variances are computed
from independently drawn samples from two normally distributed populations.
EXAMPLE 11.6
Students of two sections of a statistics course took a common final examination. A
professor examines the variability in scores between the two sections. Random sam
ples of n1 = 11 and n2 = 16 yield sample variances of s21 = 182.25 and s22= 457.96.
Construct a 95% confidence interval for the ratio of the population variances.
SOLUTION: In order to construct a 95% confidence interval for the ratio of the
population variances, we determine The degrees of
freedom in the numerator are d f 1 = n 1 − 1 = 11 − 1 = 10 and the degrees ot
freedom in the denominator are df2 = n2 − 1 = 16 − 1 = 15. From the F table and
given α = .05, we find
Therefore, a 95% confidence interval for the ratio of the population variances
ranges from 0.13 to 1.40. In other words, the variance of scores in the first section is
between 13% to 140% of the variance of scores in the second section.
As we have done in earlier chapters, we will be able to use this confidence interval to
conduct a two-tailed hypothesis test.
A two-tailed test determines whether the two population variances are different. As
noted earlier, the condition σ 2 1 = σ 2 2 is equivalent to σ2 1/σ2 2 = 1. A right-tailed test
examines whether σ21 is greater than σ22, whereas a left-tailed test examines whether σ21
is less than σ22 .
EXAMPLE 1 1.7
Let’s revisit Example 11.6.
a. Specify the competing hypotheses in order to determine whether or not the
variances in the two statistics sections differ.
b. Using the 95% confidence interval, what is the conclusion to the test?
SOLUTION:
a. Since we want to determine if the variances differ between the two sections,
we formulate a two-tailed hypothesis test as
H0: σ 2 1 / σ 2 2 = 1
HA: σ 2 1 / σ 2 2 ≠ 1
b. We calculated the 95% confidence interval for the ratio of the two vari
ances ranging from 0.13 to 1.40. We note that this interval contains the
value one; thus, we do not reject H 0. The sample data do not suggest that
the variances between the two statistics sections differ at the 5% signifi
cance level.
Now we use the four-step procedure outlined in Chapter 9 to implement one- or two-
tailed hypothesis tests. We use the ratio of the values of the sample variances s21/S22 to
conduct hypothesis tests regarding the ratio of the population variances σ 2 1 /σ 2 2 . The result
ing test is valid if the sample variances are computed from independently drawn
samples from normally distributed populations.
T E S T S T A T I S T I C F OR σ 2 1 /σ 2 2
The value of the test statistic for the hypothesis test of the ratio of two population
variances σ21/σ2 2 is computed as
We should point out that a left-tailed test can easily be converted into a right-tailed test
by interchanging the variances of the two populations. For instance, we can convert H0:
σ 2 1 /σ 2 2 ≥ 1 versus HA: σ 2 1 / σ 2 2 < 1 into H0: σ22/σ21 ≤ 1 versus HA: σ 2 2 / σ 2 1 > 1.
P L A C I N G T HE LARGER S A M PL E V A R I A N C E IN THE N U M E R A T O R
It is preferable to place the larger sample variance in the numerator of the sta
tistic. The resulting value allows us to focus only on the right tail of the distribution.
In other words, we define the hypotheses such that the resulting test statistic is computed
as s21/s22 when s21 > s22 and as s22/s21 when s 2 2 / s 2 1 ; the degrees of freedom are adjusted
accordingly. This saves us the additional work required to find the probability under the
left tail of the distribution.
E X A M P L E 1 1 .8
Let’s again visit the case introduced at the beginning of this chapter. Investment
counselor Rebecca Johnson wonders if the Metals fund is significantly riskier than
the Income fund. We assume that returns are normally distributed to implement the
test at the 5% significance level using the critical value approach. For reference, we
repeat the sample descriptive measures of the two funds:
SOLUTION: We define the population variance as the measure of risk and let
σ21 and σ22 denote the population variances of the Metals and the Income funds,
respectively. Since we wish to determine whether the variance of the Metals fund
is greater than that of the Income fund, we specify the competing hypotheses as
H0: σ21/σ22 ≤ 1
HA: σ21/σ22 > 1
Note that this specification is appropriate since s 1 = 37.13% is greater than
s2 = 11.07%. Had s2 been greater, we would have specified the hypotheses in terms
of σ22/σ21 instead of σ21/σ22.
The F test is valid because the underlying populations are assumed to be nor
mally distributed. We compute the value of the test statistic as
We reject the null hypothesis because the value of the test statistic falls in the
rejection region (F(9,9) = 11.25 exceeds the critical value F0.05,(9,9) = 3.18). At the 5%
significance level, the variance of the Metals fund is significantly greater than the
variance of the Income fund. Therefore, we can conclude that the Metals fund is
riskier than the Income fund.
Excel returns the probability of 0.0007. Again, using the p -value approach and a signifi
cance level of 5%, we reject the null hypothesis and conclude that the variance of the
Metals fund is greater than the variance of the Income fund.
TABLE 11.4 Annual Total Returns (in percent) for Metals and Income Funds
FILE Year Metals income
2000 − 7.34 4.07
2001 18.33 6.52
2002 33.35 9.38
2003 59.45 18.62
2004 8.09 9.44
2005 43.79 3.12
2006 34.30 8.15
2007 36.13 5.44
2008 − 56.02 − 11.37
2009 76.46 31.77
Source: w w w .finance.yahoo.com
A. Open the Fund Returns data from the text website into an Excel spreadsheet.
B. From the menu choose Formulas > Insert Function > F.TEST.
C. See Figure 11.9. In the FTEST dialog box, we select the data for the Metals fund for
A rrayl and then select the data for the Income fund for Array2.
D. Click OK.
Excel returns 0.0013. This is the p -value for a two-tailed test. Since we conducted a
one-tailed test, we divide this value by two (0.0013/2 = 0.0007), arriving at the p -value
that we found earlier.
FIGURE 1 1 .9 F.TEST dialog box
S Y N O P S I S OF I N T R O D U C T O R Y C A S E
E X E R C IS E S 11.2
S a m p l e 1: s 21 = 2 2 0 , a n d n 1 = 2 0
S a m p l e 2: s 22 = 1 9 6 , a n d n 2 = 15
a. C o n s t r u c t a 9 5 % in ter v a l e s t i m a t e o f t h e ra tio o f t h e b. C a lc u la te t h e a p p r o p r i a t e t e s t s ta tis tic .
p o p u l a t i o n v a r ia n c e s . c. U s i n g t h e critical v a l u e a p p r o a c h , d e t e r m i n e t h e d e c i s i o n
b. U s i n g t h e c o m p u t e d c o n f i d e n c e interval, t e s t if t h e ratio r ule a t t h e 5% s i g n i f i c a n c e le vel.
o f t h e p o p u l a t i o n v a r ia n c e s dif fe rs f r o m o n e a t t h e 5% d. Can y o u c o n c l u d e t h a t t h e v a r ia n c e s are d if fe r e n t ? G iv e n
s i g n i f i c a n c e le v e l. Explain . t h a t all o t h e r criteria a r e s a tis fie d , s h o u l d t h e c o m p a n y
2 3 . C o n s id e r t h e f o l l o w i n g c o m p e t i n g h y p o t h e s e s a n d r e l e v a n t a d o p t th e n e w tech n o lo g y ?
s u m m a r y s ta tistic s: 2 7 . T w o b a s k e t b a l l p la y e r s o n a s c h o o l t e a m a r e w o r k i n g hard
approach. differ in c o n s i s t e n c y .
b. U s e t h e critical v a l u e a p p r o a c h t o t e s t t h e c o a c h 's c la im
2 4 . C o n s id e r t h e f o l l o w i n g c o m p e t i n g h y p o t h e s e s a n d r e l e v a n t
a t α = 0.05.
s u m m a r y statistic s:
2 8 . T h e f o l l o w i n g t a b l e s h o w s t h e a n n u a l r e t u r n s (in p e r c e n t ) for
H0: σ 2 1 / σ 2 2 ≤ 1
Fidelity's E le ctro n ic a n d U tilitie s f u n d s .
HA: σ 2 1 / σ 2 2 > 1
Year E le c tro n ic U tilit ie s
S a m p l e 1 : s 21 = 9 3 5 a n d n 1 = 1 4
2005 13.23 9.3 6
S a m p l e 2: s22= 8 1 2 a n d n 2 = 11
2006 1.97 3 2 .3 3
U s e t h e critical v a l u e a p p r o a c h t o c o n d u c t th is h y p o t h e s i s
2007 2 .7 7 2 1 .0 3
t e s t a t t h e 5% s i g n i f i c a n c e le v e l. S t a t e y o u r a s s u m p t i o n s .
2008 − 5 0 .0 0 − 35.21
2 5 . C o n s id e r t h e f o l l o w i n g c o m p e t i n g h y p o t h e s e s a n d r e l e v a n t
2009 8 1 .6 5 14.71
s u m m a r y s tatistic s:
S ource : w w w . fin a n c e . y a h o o .c o m .
H0: σ 2 1 / σ 2 2 ≥ 1
U s i n g t h e critical v a l u e a p p r o a c h , t e s t if t h e p o p u l a t i o n
HA: σ 2 1 / σ 2 2 < 1
v a r i a n c e s dif fer a t t h e 5 % s i g n i f i c a n c e le v e l. S t a t e y o u r
S a m p l e 1 : s 21 = 1 , 3 1 5 a n d n 1 = 1 7
assu m p tio n s.
S a m p l e 2 : s21= 1 , 5 2 3 a n d n 2 = 1 9
2 9 . ( U s e Excel) Nike's t o t a l r e v e n u e s (in m il lio n s o f $) for t h e
C o n d u c t t h is le f t - t a ile d h y p o t h e s i s t e s t a t t h e 5% s ig n i f i c a n c e A sian a n d Latin A m e r ic a n r e g i o n s for t h e y e a r s 2 0 0 5 t h r o u g h
le v e l. You m a y w a n t t o first c o n v e r t t h e a b o v e le f t - t a ile d t e s t 2 0 0 9 a r e a s fo ll o w s :
in t o a r ig h t- ta ile d t e s t b y s w i t c h i n g t h e t w o v a r ia n c e s . S t a te
2005 2006 2007 2008 2009
your a ssu m p tio n s.
Asia 1,897 2 ,0 5 4 2 ,2 9 6 2,888 3,3 2 2
2 6 . A firm h a s j u s t d e v e l o p e d a n e w c o s t - r e d u c i n g t e c h n o l o g y
a. S p e c i f y t h e c o m p e t i n g h y p o t h e s i s in o r d e r t o t e s t
fo r p r o d u c i n g a c e r ta in r e p l a c e m e n t p a r t for a u t o m o b i l e s .
w h e t h e r t h e v a r ia n c e in r e v e n u e s is g r e a t e r in Asia t h a n
S i n c e a r e p l a c e m e n t p a r t m u s t b e p r o d u c e d w it h in c l o s e
in Latin A m e r ic a .
s p e c i f i c a t i o n s in o r d e r for it t o b e a c c e p t a b l e t o c u s t o m e r s ,
b. U s e Excel t o c a l c u l a t e d e s c r i p t i v e s ta tis tic s a n d t h e
th e n e w te c h n o lo g y 's sp ecifica tio n s m u st n o t d e v ia te
r e l e v a n t t e s t statistic .
d r a s tic a lly f r o m t h e o l d e r v e r s io n . S u p p o s e t h e s a m p l e
c. U s e Excel's F.DIST f u n c t i o n t o c a lc u la t e t h e p - v a l u e .
v a r ia n c e fo r 1 5 p a r ts p r o d u c e d u s i n g t h e o l d e r v e r s io n is
d. At α = 0 .01 w h a t is y o u r c o n c l u s i o n ?
s 21 = 0 . 3 5 , w h i l e t h e s a m p l e v a r ia n c e fo r 15 p a r ts p r o d u c e d
u s i n g t h e n e w t e c h n o l o g y is s2
2 = 0 .4 8 . A ss u m e th a t th e t w o 3 0 . F I L E ( U s e Excel) T h e m o n t h l y c l o s i n g s t o c k p r ic e s ( r o u n d e d
a. D e v e lo p t h e h y p o th e s e s to te s t w h e th e r t h e p o p u la tio n t h e f o l l o w i n g t a b l e . T h e d a t a a r e a ls o a v a ila b le o n t h e t e x t
v a r ia n c e s differ. w e b s i t e , l a b e l e d M o n t h ly S to c k Prices.
P anera Bread Co. a r e s e a l e d fo r s h i p m e n t , a n d r e p o r t t h e t e m p e r a t u r e s in
M on th S tarb u ck s Corp.
d e g r e e s C e lsiu s. W e a s s u m e t h a t t h e t w o s a m p l e s a r e d r a w n
January 2010 $22 $71
i n d e p e n d e n t l y f r o m n o r m a ll y d is t r ib u t e d p o p u l a t i o n s . A
February 2010 23 73
p o r t i o n o f t h e d a t a is s h o w n b e l o w ; t h e c o m p l e t e d a t a ,
March 2010 24 76 la b e le d P a c k a g in g , are a vailab le o n t h e te x t w e b s ite .
April 2010 26 78
26 81 N ew P ackage O ld P a c k a g e
May 2010
June 2010 24 75 3.98 5 .7 9
4 .9 9 6 .4 2
Source: w w w .fin a n c e .y a h o o .c o m .
⋮ ⋮
a. S t a t e t h e null a n d t h e a lt e r n a t i v e h y p o t h e s e s in o r d e r t o
4.95 5.95
d e t e r m i n e if t h e v a r ia n c e o f p r ic e differs for t h e t w o firms.
b. W h a t a s s u m p t i o n r e g a r d i n g t h e p o p u l a t i o n is n e c e s s a r y a. S ta te t h e a p p rop riate h y p o t h e s e s t o t e s t w h e t h e r or n o t
t o i m p l e m e n t t h is s t e p ? t h e n e w p a c k a g i n g m a t e r ia l r e d u c e s t h e v a r ia tio n o f
c. U s e Excel's F.TEST f u n c t i o n t o c a l c u l a t e t h e p -v a lu e . t e m p e r a t u r e s in t h e b o x .
d. At α = 0 . 0 5 w h a t is y o u r c o n c l u s i o n ? b. U s e Excel's F.TEST f u n c t i o n t o c a l c u l a t e t h e p -v a lu e .
c. M a k e a c o n c l u s i o n a t t h e 5 % s i g n i f i c a n c e le v e l.
3 1 . F I L E ( U s e Excel) A v a r ie ty o f p a c k a g i n g s o l u t i o n s e x i s t for
p ro d u cts th a t m u st b e k ept w ith in a specific te m p er a tu re 32. F I L E A ccess th e data la b eled R e n ta ls o n th e te x t w e b s ite th at
is tr y in g o u t a n e w p a c k a g i n g m a t e r ia l t h a t m i g h t r e d u c e Davis, California, is k n o w n t o h a v e h i g h e r r e n ts th a n A n n
t h e v a r ia tio n o f t e m p e r a t u r e s in t h e b o x . It is b e l i e v e d t h a t Arbor, M ic h ig a n ; h o w e v e r , it is n o t c le a r if it a l s o h a s h ig h e r
1 6 b o x e s o f n e w a n d o ld p a c k a g e s , 2 4 h o u r s a ft e r t h e y Arbor, M ic h ig a n . S t a t e y o u r a s s u m p t i o n s clearly.
WRITING WI TH S T A T I S T I C S
60 70
55 65
⋮ ⋮
52 65
Elizabeth would like to use the above sample information to:
1. Determine, at the 5% significance level, whether the standard deviation on the
55-mph highway exceeds 5 mph.
2. Determine, at the 5% significance level, whether the variability on the 55-mph
highway is more than the variability on the 65-mph highway.
Increasing gas prices are prompting conservationists to lobby for a return to the federal Sample
55-mile-per-hour (mph) speed limit on America’s highways. Advocates point to potential
money and fuel savings, noting that fuel efficiency worsens at speeds above 60 mph. It is Report—
not clear, however, if the return to 55 mph will increase traffic safety. Many believe that
traffic safety is based on the variability of the speed rather than the average speed that Traffic Safety
people are driving— the more variation in speed, the more dangerous the roads.
In this report the variability of speeds on two highways is compared. The sample and the
consists of the speeds of 40 cars recorded on a highway with a 55-mph speed limit
(Highway 1) and the speeds of 40 cars recorded on a highway with a 65-mph speed limit Variation
(Highway 2). Table 11.A shows the most relevant descriptive measures for the analysis.
TABLE 1 1 .A Summary Measures for Highway 1 and Highway 2
in Speed
H ig h w a y 1 H ig h w a y 2
(5 5 -m p h s p e e d lim it) (6 5 -m p h s p e e d lim it)
Mean 56.60 66.00
Standard deviation 6.98 3.00
Number of cars 40 40
While it is true that cars travel at a slower speed, on average, on Highway 1 (56.60 mph <
66.00 mph), the variability of speeds is greater on Highway 1 as measured by standard
deviation (6.98 mph > 3.00 mph).
Two hypothesis tests are conducted. The first test examines whether or not the standard
deviation on Highway 1 is greater than 5 mph at the 5% significance level, or alterna
tively σ 2 > 52. The second test analyzes whether the standard deviation on Highway 1 is
significantly greater than the standard deviation on Highway 2, or alternatively σ21/σ22>
1. The results of the tests are summarized in Table 11.B.
TABLE 1 1 .B C om peting H yp o th e se s, Test Statistics, and p -Values
H y p o th e s e s Test S ta tistic p -V a lu e
H0: σ2 ≤ 52
HA: σ2 > 52
0.00
H0: σ21/σ22 ≤ 1
0.00
HA: σ21/σ22 > 1
When testing whether or not the standard deviation is greater than 5 mph on Highway 1, a
test statistic of 76.00 is obtained. Given its p -value of 0.00, the null hypothesis regarding
the population variance is rejected at any reasonable level of significance. In other words,
the sample data suggest that the standard deviation is significantly greater than 5 mph on
Highway 1. With a test statistic of 5.41 and a corresponding p -value of 0.00, the second
hypothesis test reveals that the variance for Highway 1 is significantly greater than the
variance for Highway 2.
American drivers love to drive fast, which explains why safety advocates and conser
vationists are losing the long-running debate over lowering highway speed limits. While
a 55-mph limit will save fuel and reduce greenhouse emissions, it is still an open question
as to whether it will also enhance safety. If traffic safety is based on the variability of the
speeds that people are driving rather than the average speed, then the data suggest that a
return to a federal 55-mph speed limit may not necessarily enhance safety.
C o n cep tu al Review
LO 1 1 .1 D isc u ss fe a tu r e s o f t h e χ 2 d istrib u tio n .
The χ 2 d istr ib u tio n is characterized by a family of distributions, where each dis
tribution depends on its particular degrees of freedom df. It is common, therefore,
to refer to it as the χ2df distribution. It is positively skewed with values ranging
from zero to infinity. As the d f grow larger, the χ2df distribution tends to the normal
distribution.
2004 7.11 39. F I L E (Use Excel) Barbara Dwyer, the manager at Lux Hotel,
makes every effort to ensure that customers attempting to
2005 – 1.75
make phone reservations do not have to wait too long to
2006 13.33
speak with a reservation specialist. Since the hotel accepts
2007 0.01
phone reservations 24 hours a day, Barbara is especially
2008 – 61.20 interested in maintaining consistency in service. Barbara
2009 122.28 wants to determine if the variance of wait time in the early
Source: h ttp ://b iz .y a h o o .c o m . morning shift differs from that in the late morning shift. She
uses the following independently drawn samples o f wait
a. State the null and the alternative hypotheses in time for phone reservations for both shifts for the analysis.
order to test whether the standard deviation is greater The data can be found on the text website and are labeled
than 35%. W a it T im es. Assume that wait times are normally distributed.
Shift W ait Tim e (in seconds) 40. FILE (U se Excel) Ad idas r e v e n u e s (in millions o f €) in Asia
Early Morning
67 48 83 66 a n d Latin America for th e years 2 0 0 5 th r o u g h 2 0 0 9 are
2
5 7 1 9
5 9
4
Shift: 12:00 a m - 57 68
6 0 6 82 83
s h o w n in t h e a c c o m p a n y i n g table; t h e data, l a b eled Adidas
6 3 6
4 Revenues, are also available on the text website.
6:00 am 37 41 60 41 8 7 53 66 69
98 100 122 108 100 123 102 2005 2006 2007 2008 2 009
Late M orning 90
Asia 1,523 2,020 2,254 2,662 2,614
Shift: 6:00 a m - 125 121 120 128 94 1 2 8 113
3
2
1
12:00 pm Latin America 319 499 657 893 1,006
116 104 96 111 107 105 113 106
S OURCE: Adidas Online Annual Reports.
a. Specify t h e ap propriate h y p o t h e s e s t o te s t if t h e va riance
o f w a i t t i m e in t h e e a r l y m o r n i n g s h ift d if fe rs f r o m t h a t in a. Specify the competing hypothesis in order to test
c. D o e s t h e variance o f w ait t i m e in t h e early m orn ing shift c. Use Excel's F.TEST function to calculate the p -value.
differ from t h a t in t h e late after n o o n shift? d. At α = 0.05 w hat is your conclusion?
CASE STUDIES
LEARNING OBJECTIVES
After reading
E R
Chi-Square Tests
LO 1 2 .1 Conduct a goodness-of-fit test for a m ultinom ial experim ent.
this chapter
you should LO 1 2 .2 Determine whether two classifications of a population are independent.
be able to:
LO 1 2.3 Conduct a goodness-of-fit test for normality.
In this chapter we focus on the χ 2 (chi-square) distribution to develop statistical tests tha t compare
observed data w ith w h at we w ould expect from a population w ith a specific distribution. Generally,
the chi-square tests are used to assess tw o types o f comparison. First, a goodness-of-fit test is co m
m only used w ith frequency data representing the various outcomes o f a qualitative variable. For
instance, we may w a nt to substantiate a claim tha t market shares in the autom otive industry have
changed dramatically over the past 10 years. Whereas a goodness-of-fit test focuses on a single qual
itative variable, a test o f independence is used to compare tw o qualitative variables. For example, we
may w a nt to determ ine w hether a person's gender influences his/her purchase o f a product. We can
also extend the goodness-of-fit test to determine w h ether it is reasonable to assume that sample
data are drawn from a normal population. Since we use the normal distribution w ith quantitative
data, we first convert the raw data into a frequency distribution, where each interval is now viewed
as a category. Finally, w e introduce the Jarque-Bera test, which allows us to test for norm ality using
the data in their raw form.
I N T R O D U C T O R Y CASE
Sportswear Brands
In C h a p t e r 4, w e p r e s e n t e d a c a se s t u d y in w h ic h A n n a b e l G o n z a le z , c h ie f r e ta il a n a ly s t a t a m a r k e t in g f ir m ,
s t u d ie d t h e r e la t io n s h ip b e t w e e n t h e b r a n d n a m e o f c o m p r e s s io n g a r m e n t s in t h e s p o r t - a p p a r e l in d u s t r y a n d
t h e a g e o f t h e c o n s u m e r . S p e c ific a lly , s h e w a n t e d t o k n o w w h e t h e r t h e a g e o f t h e c o n s u m e r in f lu e n c e d t h e b r a n d
n a m e p u r c h a s e d . S h e w o u ld lik e t o c o n d u c t a f o r m a l t e s t o f in d e p e n d e n c e b e t w e e n t h e a g e o f t h e c o n s u m e r a n d
th e b ra n d n a m e p u rc h a s e d .
H e r in it ia l f e e lin g is t h a t t h e U n d e r A r m o u r b r a n d a t t r a c t s a y o u n g e r c u s t o m e r , w h e r e a s t h e m o r e e s ta b lis h e d
c o m p a n ie s , N ik e a n d A d id a s , d r a w a n o ld e r c lie n t e le . S h e b e lie v e s t h is in f o r m a t io n is r e le v a n t t o a d v e r tis e r s a n d
r e ta ile r s in t h e s p o r t in g - g o o d s in d u s t r y , as w e ll as t o s o m e in t h e f in a n c ia l c o m m u n it y . S u p p o s e s h e c o lle c t s d a ta
o n 6 0 0 r e c e n t p u rc h a s e s in t h e c o m p r e s s io n - g e a r m a r k e t . T h e c o n t in g e n c y t a b le (c ro s s -c la s s ifie d b y a g e a n d
b r a n d n a m e ) f r o m C h a p t e r 4 is r e p r o d u c e d h e re as T a b le 12.1 f o r e a s e o f e x p o s it io n .
Brand Name
Age Group Under Armour Nike Adidas
35 years or older 54 72 78
A n n a b e l w a n t s t o u se t h e a b o v e s a m p le in f o r m a t io n to :
1. D e t e r m in e w h e t h e r t h e t w o c la s s ific a tio n s ( a g e a n d b r a n d n a m e ) a re d e p e n d e n t a t t h e 5 % s ig n if ic a n c e le v e l.
2. D is cu s s h o w t h e f in d in g s f r o m t h e t e s t f o r in d e p e n d e n c e c a n b e u s e d .
A M U L T IN O M IA L E X P E R IM E N T
• As compared from the previous day, a stockbroker records whether the price of
a stock rises, falls, or stays the same. This example has three possible outcomes
(k = 3).
• A consumer rates service at a restaurant as excellent, good, fair, or poor (k = 4).
• The admissions office records which of the six business concentrations a student
picks (k = 6).
When setting up the competing hypotheses for a multinomial experiment, we have es
sentially two choices. We can set all population proportions equal to the same specific
value, or equivalently, equal to one another. For instance, if we want to judge on the basis
of sample data whether the proportion of voters who favor four different candidates is the
same, the competing hypotheses would take the following form:
H0: p1 = p2 = p3 = p4 = 0.25
HA: Not all population proportions are equal to 0.25.
Note that the hypothesized value under the null hypothesis is 0.25 because the population
proportions must sum to one. We can also set each population proportion equal to a dif
ferent predetermined (hypothesized) value. Suppose we want to determine whether 40%
of the voters favor Candidate 1, 30% favor Candidate 2, 20% favor Candidate 3, and 10%
favor Candidate 4. The competing hypotheses are formulated as
H0: p 1 = 0.40, p 2 = 0.30, p 3 = 0.20, and p 4 = 0.10
HA: At least one of the proportions is different from its hypothesized value.
When conducting a statistical test on a multinomial experiment, we take a random sample
and determine whether the sample proportions are close enough to the hypothesized pop
ulation proportions under the null hypothesis. For this reason, this type of test is called a
goodness-of-fit test. Under the usual assumption that the null hypothesis is true, we de
rive the expected frequencies of the outcomes of a multinomial experiment and compare
them with observed frequencies. The objective is to determine whether we can reject the
null hypothesis in favor of the alternative hypothesis. To see how to conduct a goodness-
of-fit test, consider the following example.
One year ago, the management at a restaurant chain surveyed its patrons to determine
whether changes should be made in the menu. One question on the survey asked patrons
to rate the quality of the restaurant’s entrées. The percentages of the patrons responding
Excellent, Good, Fair, or Poor are listed in the following table:
Based on responses to the overall survey, management decided to revamp the menu.
Recently, the same question concerning the quality of entrées was asked of a random
sample of 250 patrons. Their responses are shown below:
At the 5% significance level, we want to determine whether there has been any change in
the population proportions calculated one year ago.
Since we want to determine whether the responses of the 250 patrons are consistent
with the earlier proportions, we let the earlier population proportions denote the hypoth
esized probabilities for the test. Thus, we use p 1, p 2, p 3, and p 4 to denote the population
proportions of those that responded Excellent, Good, Fair, or Poor, respectively, and con
struct the following competing hypotheses.
H0: p 1 = 0.15, p 2 = 0.30, p 3 = 0.45, and p 4 = 0.10
HA: At least one of the proportions is different from its hypothesized value.
The first step in calculating the test statistic is to calculate an expected frequency for
each category. That is, we need to estimate the frequencies that we would expect to get
if the null hypothesis is true. In general, in order to calculate the expected frequency
ei for category i we multiply the sample size n by the respective hypothesized value
of the population proportion p i. For example, consider the outcome Excellent. If H0
is true, then we expect that 15% (p 1 = 0.15) of 250 patrons will find the quality of
entrées to be excellent. Therefore, the expected frequency of Excellent responses is
37.5 (= 250 × 0.15), whereas the corresponding observed frequency is 46. Expected
frequencies for other responses are found similarly. Ultimately, when computing the
value of the test statistic, we compare these expected frequencies to the frequencies we
actually observe.
G O O D N E S S - O F - F IT TEST STATISTIC
For a multinomial experiment with k categories, the test statistic follows the χ2df
distribution with k − 1 degrees of freedom. The value of the test statistic is calcu
lated as
where oi and ei = n × p i are the observed and expected frequency in the ith category,
respectively.
Note: The test is valid so long as the expected frequencies in each category are five
or more.
Table 12.2 shows the expected frequency ei for each category. The condition that each
expected frequency ei must equal five or more is satisfied here. As we will see shortly,
sometimes it is necessary to combine data from two or more categories to achieve this
result.
As a check on the calculations, the sum of the expected frequencies must equal the
sample size n, which in this example equals 250. Once the expected frequencies are esti
mated, we are ready to compute the value of the test statistic.
The χ2df statistic measures how much the observed frequencies vary (or differ) from
the expected frequencies. In particular, χ2df is computed as the sum of the standardized
squared deviations. The smallest value that χ2df can assume is zero— this occurs when
each observed frequency equals its expected frequency. Rejection of the null hypothesis
occurs when χ2df is significantly greater than zero. As a result, these tests of hypotheses
regarding multiple population proportions (p 1, p 2, p 3, . . .) are always implemented as
right-tailed tests. However, since the alternative hypothesis simply states that at least one
of the proportions is not equal to its specified value in the null hypothesis, rejection of the
null hypothesis does not indicate which proportion differs.
In this example, there are four categories (k = 4), so the degrees of freedom are equal
to three. The value of the test statistic is calculated as
Since a goodness-of-fit test is a right-tailed test, the critical value with α = 0.05 and
d f = 3 is found from th e χ2 (chi-square) table as χ2α,df = χ20.05,3 = 7.815; we show a portion
of the χ2 table in Table 12.3 as a refresher. Hence, our decision rule is to reject H0 if
3> 7.815.
χ2
Given that χ23 = 6.52 < 7.815, we decide not to reject H0. We cannot conclude that the
proportions differ from the ones from one year ago at the 5% significance level. Manage
ment may find this news disappointing in that the goal of the menu change was to improve
TA B LE 12.3 Portion of the χ2
Stable
Area in Upper Tail, a
df 0.995 0.990 0.975 0.950 0.900 0.100 0.050 0.025 0.010 0.005
1 0.000 0.000 0.001 0.004 0.016 2.706 3.841 5.024 6.635 7.879
2 0.010 0.020 0.051 0.103 0.211 4.605 5.991 7.378 9.210 10.597
3 0.072 0.115 0.216 0.352 0.584 6.251 7.815 9.348 11.345 12.838
customer satisfaction. Responses to other questions on the survey may shed more light on
whether the goals of the menu change met or fell short of expectations.
Excel returns the probability of 0.0889, which is equivalent to the p-value. Again, given a
significance level of 5%, we are unable to reject the null hypothesis.
EXAMPLE 12.1
Table 12.4 lists the market share in 2010 of the five firms that manufacture a particu
lar product. A marketing analyst wonders whether the market shares have changed
since 2010. He surveys 200 customers. The last column of Table 12.4 shows the
number of customers who recently purchased the product at each firm.
TA B LE 12.4 Market Share of Five Firms
M arket Share Num ber of
Firm in 2010 Recent Customers
1 0.40 70
2 0.32 60
3 0.24 54
4 0.02 10
5 0.02 6
a. Specify the competing hypotheses to test whether the market shares have
changed since 2010.
b. Calculate the value of the test statistic.
c. At the 5% significance level, what is the critical value?
d. Have the market shares changed since 2010?
S O LU TIO N :
a. Let p i denote the market share for the ith firm. In order to test whether the
market shares have changed since 2010 we initially set up the competing
hypotheses as
H0: p 1 = 0.40, p 2 = 0.32, p 3 = 0.24, p 4 = 0.02, and p 5 = 0.02
HA: At least one of the market shares is different from its hypothesized value.
We note that expected frequencies for firms 4 and 5 are less than five. The test
is valid so long as the expected frequencies in each category are five or more.
In order to achieve this result, we combine the expected frequencies for firms
4 and 5 to obtain a combined frequency of eight (e4 + e5 = 8). We could have
made other combinations, say e4 with e 1 and e5 with e2, but we preferred to
maintain a category for the less dominant firms. After making this combina
tion, we now re-specify the competing hypotheses as
H 0 : p 1 = 0 .4 0 , p 2 = 0 .3 2 , p 3
= A
HAt
: 0least
. 2one4 of, the market
a n dsharesp is4different
= from
0 . its
0 hypothesized
4 value.
c. Given α = 0.05, we find the critical value as χ20.05,3 = 7.815; thus, the decision
rule is to reject H0 if χ23 > 7.815.
d. Since 10.25 is greater than 7.815, we reject the null hypothesis and conclude
that at least one of the proportions (market shares) differs from its 2010 value.
As mentioned earlier, one limitation of this type of chi-square test is that we can
not tell which proportion differs. However, given the divergence between the
observed and expected frequencies for the less dominant firms, it appears that they
may be making some headway in this industry. Further testing can be conducted
to see if this is the case.
EXERCISES 12. 1
In this section, we use the data in a contingency table to conduct a hypothesis test that
determines whether the classifications depend upon one another. Whereas a goodness-
of-fit test examines a single qualitative variable, a test of independence— also called a
chi-square test of a contingency table— analyzes the relationship between two qualita
tive variables. Many examples of the use of this test arise, especially in marketing, bio
medical research, and the courts of law. For instance, a retailer may be trying to determine
whether there is a relationship between the age of its clientele and where it chooses to ad
vertise. Doctors might want to investigate whether or not losing weight through stomach
surgery can extend the lives of severely obese patients. Or, one party in a discrimination
lawsuit may be trying to show that gender and promotion are not independent events. All
of these examples lend themselves to applications of the hypothesis test discussed in this
section.
In the introductory case study, we are presented with a contingency table cross
classified by age and brand name. Specifically, we want to determine whether or not the
age of a consumer influences his/her decision to buy a garment from Under Armour, Nike,
or Adidas. We will conduct this test at the 5% significance level.
In general, the competing hypotheses for a statistical test of independence are formu
lated such that rejecting the null hypothesis leads to the conclusion that the two categories
are dependent. Formally,
H0: The two classifications are independent.
HA: The two classifications are dependent.
Since the criteria upon which we classify the data are brand name and age, we write the
competing hypotheses as
H0: Age and brand name are independent.
HA: Age and brand name are dependent.
Table 12.5 reproduces Table 12.1 of the introductory case. The category “age” has two pos
sible outcomes: (1) under 35 years and (2) 35 years or older. The category “brand name”
has three possible outcomes: (1) Under Armour, (2) Nike, and (3) Adidas. Each cell in this
table represents an observed frequency oij where the subscript ij refers to the ith row and the
j th column. Thus, o 13 refers to the cell in the first row and the third column. Here, o 13 = 90,
or equivalently, 90 customers under 35 years of age purchased an Adidas product.
Note that we have not specified a value under the null hypothesis. However, we will use
the independence assumption postulated under the null hypothesis to derive an expected
frequency for each cell from the sample data. In other words, we first estimate values in
Table 12.5 as if no relationship exists between the age of a consumer and the brand name
of the clothing purchased. Then we will compare these expected frequencies with the
observed values to compute the value of the test statistic.
We can now calculate each cell probability by recalling from Chapter 4 the multiplication
rule for independent events. That is, if two events are independent, say events A 1 and B 1
(our assumption under the null hypothesis), then their joint probability is
Multiplying this joint probability by the sample size yields the expected frequency for
cell e11:
e11 = 600(0.2508) = 150.48.
We use the following general formula to calculate the expected frequencies for each
cell in a contingency table:
Table 12.7 shows the expected frequency eij of each cell. In order to satisfy subsequent
assumptions, each expected frequency eij must equal five or more. This condition is satis
fied here. As we saw in Example 12.1, it may be necessary to combine two or more rows
or columns to achieve this result in other applications.
When conducting a test of independence, we apply the chi-square test statistic f2.χd
Analogous to the discussion in Section 12.1, χ2df measures how much the observed
frequencies vary (or differ) from the expected frequencies. The smallest value that χ2df can
assume is zero— this occurs when each observed frequency equals its expected frequency.
Thus, a test of independence is also implemented as right-tailed test.
where oij and eij are the observed and the expected frequencies in the ith row and the
j th column, respectively.
Note: This test is valid only when the expected frequencies in each cell are five or
more.
With two rows and three columns in the contingency table, degrees of freedom are calcu
lated as d f = (r − l)(c − 1) = (2 − 1)(3 − 1) = 2. We apply the formula to compute the
value of the test statistic as
Given a significance level of 5% and d f = 2, we find the critical value as χ2α,df = χ20.05,2 = 5.991.
Hence, the decision rule is to reject H0 if χ22 > 5.991. Since χ22 = 22.53 > 5.991, we reject
H0. At the 5% significance level, we conclude that the two qualitative variables are depen
dent; that is, there is a relationship between the age of a consumer and the brand name of
the apparel purchased.
As usual, we can conduct the above hypothesis test by using the p-value approach to
hypothesis testing rather than the critical value approach, where the p -value is derived
as P(χ22 ≥ 22.53). As discussed earlier, we can calculate an exact p -value with Excel by
selecting Formulas > Insert Function > CHISQ.DIST.RT. By inputting X = 22.53
and Deg_freedom = 2, Excel returns an almost zero p -value of 1.28 X 10−5. Again, given
a significance level of 5%, we reject the null hypothesis and conclude that age and brand
name are not independent of one another.
EXAMPLE 12.2
A recent study of gender preferences among car shoppers found that men and
women equally favor economy cars (www.cargurus.com, February 14, 2011). A
marketing analyst doubts these results. He believes that a person’s gender influences
whether or not he/she purchases an economy car. He collects data on 400 recent
car purchases cross-classified by gender and type of car (economy car versus non
economy car). The results are shown in Table 12.8. At the 10% significance level,
determine whether the sample data support the marketing analyst’s claim.
With two rows (r = 2) and two columns (c = 2) in the contingency table, we com
pute degrees of freedom as d f = (r − l)(c − 1) = (2 − 1)(2 − 1) = 1. The value
of the test statistic is
Given α = 0.10, we find the critical value a s χ20.10,1 = 2.706; Since χ21 = 0.55 < 2.706,
we do not reject the null hypothesis. The sample data do not support the market
ing analyst’s claim that gender differences exist with respect to the purchase of an
economy car.
S Y N O P S I S OF I N T R O D U C T O R Y C A S E
U n d e r A r m o u r p io n e e r e d c lo t h in g in t h e c o m p r e s s io n - g e a r m a r k e t. C o m p r e s s io n g a r
m e n t s a re m e a n t t o k e e p m o is tu r e a w a y f r o m a w e a re r's b o d y d u r in g a t h le tic a c tiv itie s
in w a r m a n d c o o l w e a th e r . U n d e r A r m o u r has e x p e r ie n c e d e x p o n e n t ia l g r o w t h sin c e
t h e f ir m w e n t p u b lic in N o v e m b e r 2 0 0 5 (USA Today, J u n e 1 6 , 2 0 1 0 ); h o w e v e r, N ik e a n d
A d id a s h a v e a g g re s s iv e ly e n te r e d t h e c o m p r e s s io n - g e a r m a r k e t as w e ll. A n a n a ly s is
is c o n d u c t e d t o e x a m in e w h e t h e r t h e a g e o f t h e c u s t o m e r m a tte r s w h e n m a k in g a
p u rc h a s e in t h e c o m p r e s s io n - g e a r m a r k e t. T h is in f o r m a t io n is r e le v a n t n o t o n ly f o r
U n d e r A r m o u r a n d h o w t h e f ir m m a y fo c u s its a d v e r t is in g e ffo rts , b u t a ls o t o c o m p e t i
to r s a n d re ta ile rs in t h is m a r k e t. D a ta w e r e c o lle c te d o n 6 0 0 re c e n t p u rc h a s e s in t h e
c o m p r e s s io n - g e a r m a r k e t; t h e d a ta w e r e t h e n c ro s s -c la s s ifie d b y a g e g r o u p a n d b r a n d
n a m e . A t e s t o f in d e p e n d e n c e w a s c o n d u c t e d a t t h e 5 % s ig n ific a n c e le v e l. T h e re s u lts
s u g g e s t t h a t a c u s to m e r 's a g e a n d t h e b r a n d n a m e p u rc h a s e d a re d e p e n d e n t o n o n e
a n o th e r . G iv e n t h a t a g e in flu e n c e s t h e b r a n d n a m e p u rc h a s e d , it is n o t s u r p r is in g t h a t
U n d e r A r m o u r r e c e n tly s ig n e d N F L q u a r t e r b a c k T o m B ra d y (c n b c .c o m , O c t o b e r 6 ,
2 0 1 0 ) t o e n d o r s e its p ro d u c ts , a m o v e lik e ly t o a t t r a c t a y o u n g e r c o n s u m e r . B ra d y h a d
s p e n t m o s t o f h is c a re e r w i t h N ik e b e fo r e b r e a k in g a w a y t o g o w it h U n d e r A r m o u r .
EXERCISES 12.2
12. Suppose you are c o n d u c tin g a te s t o f independence.
Concepts
Specify th e critical value u n d e r th e fo llo w in g
11. Suppose you are c o n d u c tin g a te s t o f independence. Specify
scenarios:
th e critical value u n d e r th e fo llo w in g scenarios:
a. rows = 5, co lu m n s = 2, and α = 0.025.
a. rows = 3, colu m n s = 3, and α = 0.10.
b. rows = 3, co lu m n s = 5, and α = 0.01.
b. rows = 4, co lu m n s = 5, a nd α = 0.05.
13. Given th e fo llo w in g c o n tin g e n c y table, c o n d u c t a te st o f a. Specify th e c o m p e tin g hypotheses to d e term in e
inde pe n d e n ce a t th e 5% significance level using (a) th e w h e th e r co lo r preference depends on gender.
critical value approach and (b) th e p -value approach. b. C om pute th e value o f th e te st statistic.
c. Specify th e critical value at th e 1% significance level.
Category 1
d. Does yo u r conclusion suggest th a t th e com p a n y should
Category 2 1 2
ta rg e t a d ve rtise m en ts d iffe re n tly fo r males versus
1 23 47
females? Explain.
2 32 53
17. The fo llo w in g sam ple data reflect shipm ents received by a
14. Given th e fo llo w in g c o n tin g e n c y table, c o n d u c t a te st o f large firm fro m th re e d iffe re n t vendors.
inde pe n d e n ce a t th e 1% significance level using (a) th e
p -value approach and (b) th e critical value approach. Vendor Defective Acceptable
1 14 112
Category 1
2 10 70
Category 2 1 2 3 4
15. According to an o n lin e survey by Harris Interactive fo r jo b site c. Should th e firm be concerned a b o u t th e source o f th e
CareerBuilder.com (lnform ationW eek.com , S eptem ber 27, shipm ents? Explain.
2007), m ore th a n h a lf o f IT workers say th e y have fallen asleep 18. A ccording to a 2008 survey by th e Pew Research
at w ork. S ixty-four percent o f g o v e rn m e n t workers ad m itte d Center, peo ple in China are h ig h ly satisfied w ith
to fa llin g asleep on th e jo b . Assume th a t th e fo llo w in g th e ir roaring e c o n om y and th e d irectio n o f th e ir nation
co n tin g e n c y ta b le is representative o f th e survey results. (USA Today, July 2 2 , 2008). Eighty-six percent o f those
w h o were surveyed expressed positive view s o f th e
Job Category
w a y China is progressing and described th e econom ic
Governm ent s itu a tio n as good. A p o litica l analyst w a n ts to kn o w if
Slept on the Job? IT Professional Professional th is o p tim is m a m o n g th e Chinese depends on age. In
Yes 155 256 an in d e p e n d e n t survey o f 280 Chinese residents, th e
respondents are asked h o w h a p py th e y are w ith th e
No 145 144
d ire c tio n th a t th e ir c o u n try is ta kin g . Their responses are
ta b u la te d below.
a. Specify th e co m p e tin g hypotheses to d e te rm in e w h e th e r
sleeping on th e jo b is associated w ith jo b category.
Somewhat
b. C o m pu te th e value o f th e te st statistic. Age Very Happy Happy Not Happy
c. A p p ro x im a te th e p -value. 20 up to 40 23 50 18
d. A t th e 5% significance level, can you co nclude th a t
40 up to 60 51 38 16
sleeping on th e jo b depends on jo b category?
60 and above 19 45 20
16. A m arke t researcher fo r an a u to m o b ile c o m p a n y suspects
differences in preferred co lo r b e tw e en m ale and fem ale
a. Set up th e a p p ro p ria te hypotheses to te st th e claim th a t
buyers. A d ve rtise m e n ts ta rg e te d to d iffe re n t groups
o p tim is m regarding China's d irectio n depends on th e
should take such differences in to account, if th e y exist. The
age o f th e respondent.
researcher exam ines th e m ost recent sales in fo rm a tio n o f a
b. Use Excel to c o m p u te th e p -value fo r th e test.
pa rticu la r car th a t comes in th re e colors.
c. A t a 1% level o f significance, can w e infer th a t o p tim is m
Gender of Autom obile Buyer am o n g th e Chinese is d e p e n d e n t on age?
Color Male Female 19. A stu d y by th e Massachusetts C o m m u n ity & Banking Council
Silver 470 280 fo u n d th a t blacks, and to a lesser extent, Latinos, rem ain
largely unable to b o rro w m oney at th e same interest rate as
Black 535 285
w h ite s (The Boston Globe, February 28,2008). The fo llo w in g
Red 495 350
c o n tin g e n c y ta b le shows representative data fo r th e city
o f Boston, cross-classified by race and ty p e o f interest rate looked a t w h y Facebook users break up in a relatio n sh ip
received: (The Wall Street Journal, N ove m b e r 2 7 - 2 8 , 2010).
Type o f Interest Rate on Loan G ender
Race High Interest Rate Lower Interest Rate Reasons for Breakup Men Women
Lost Interest
28% 26%
At th e 5% significance level, do th e data indicate th a t th e
interest rate received on a loan is d e p e n d e n t on race? O ther 30% 32%
20. Founded in February 2004, Facebook is a social u tility th a t Suppose th e survey consisted o f 1,800 m en and 1,200
helps pe o ple co m m u n ic a te w ith th e ir frien d s and fam ily. In w o m e n . Use th e data to d e te rm in e w h e th e r th e reasons
ju s t six years, Facebook has acquired m ore th a n 500 m illio n fo r breakup d e pend on g e n de r at th e 1% significance level.
active users, o f w h ic h 50% log on to Facebook in any given Provide th e details.
day. In a survey o f 3,000 Facebook users, th e designers
25 38 14 65 77 110 22 18 180 52
44 17 45 99 250 78 58 35 57 45
37 58 62 44 35 78 35 82 94 58
We first use the sample data to compute the sample mean and the sample standard
deviation as
Since we want to determine whether or not the data may be looked upon as a random
sample from a population having a normal distribution, we specify this in the null hypoth
esis, along with the above sample estimates of the population mean and the population
standard deviation.
H0: Income (in $1,000s) in a given small Midwestern city follows a normal distribu
tion with mean $63.80 and standard deviation $45.78.
HA: Income (in $1,000s) in a given small Midwestern city does not follow a normal
distribution with mean $63.80 and standard deviation $45.78.
The null hypothesis implies that the underlying distribution is normal and also that the
population mean and the population standard deviation equal their estimates. In other
words, it uses μ = 63.80 and σ = 45.78. As discussed in Section 12.1, the goodness-of-
fit test for a multinomial experiment deals with a single population of qualitative data.
Since observations that follow the normal distribution are quantitative, we essentially
need to convert the data into a qualitative format. After computing the sample mean and
the sample standard deviation, we subdivide the data into non-overlapping intervals (cat
egories); in other words, we construct a frequency distribution. The first two columns of
Table 12.10 show the frequency distribution for the raw data from Table 12.9.
n = Σ o i = 50 Σp i = 1 n= Σ ei = 50
Note that we have six observations (15, 19, 19, 14, 18, and 17) that are less than 20. Other
frequencies are found similarly. Earlier, we were able to calculate expected frequencies by
multiplying the sample size n by the hypothesized probabilities (proportions) p i under the
null hypothesis. Here, we first calculate the probabilities under the assumption of a normal
distribution and then use them to calculate expected frequencies. For example, under the
null hypothesis that income is normally distributed with μ = 63.80 and σ = 45.78, we
reference the z table to find the probability that an individual’s income is less than 20, or
The third column of Table 12.10 shows these probabilities. We are then able to compute
the expected frequencies for each category as n × p i. The fourth column of Table 12.10
shows the values for the expected frequencies. As in Section 12.1, the appropriate test
statistic follows the χ2df distribution and its value is calculated as The only
difference is that the degrees of freedom are equal to the number of categories minus one,
minus the number of parameters estimated. Since we estimate two parameters— the mean
and the standard deviation— from the sample data, the degrees of freedom for the chi-
square test for normality are always k − 1 − 2 = k − 3.
G O O D N E S S - O F - F IT TEST STATISTIC FOR N O R M A L I T Y
The goodness-of-fit test statistic for normality follows the χ2df distribution with
d f = k − 3 and its value is calculated as
where oi and ei are the observed frequency and the expected frequency in the ith
category, respectively, and k is the number of class intervals (categories) of the
frequency distribution.
Note: When constructing the frequency distribution, we must ensure that the
expected frequencies for each class equal five or more. If necessary, the number
of classes can be adjusted by combining adjacent classes until this condition is
achieved.
The Jarque-Bera JB test statistic for normality follows the χ2df distribution with
d f = 2 and its value is calculated as
where n is the sample size, S is the skewness coefficient, and K is the kurtosis
coefficient.
EXAMPLE 12.3
Using the data from Table 12.9 and the Jarque-Bera test, determine whether or not
annual household income is normally distributed at the 1% significance level.
SO LU TIO N :
The competing hypotheses take the following form:
H0: S = 0 and K = 0
HA: S ≠ 0 or K ≠ 0
In order to compute the value of the test statistic, we first need to compute the skew
ness and kurtosis coefficients, S and K. We can use the SKEW and KURT functions
in Excel to approximate these values. Alternatively, we can choose Data > Data
Analysis > Descriptive Statistics whereby Excel reports approximate S and K
along with other descriptive measures. Using either method, we find that S = 2.32
and K = 6.73.
The value of the test statistic is calculated as
With a significance level of 1% and d f = 2, we find the critical value χ2df = χ20.01,2 =
9.210. Since χ22 > χ20.01,2 (139.21 > 9.210), we reject H0 and conclude that income in
this Midwestern city does not follow a normal distribution.
In the above examples, the conclusion with the Jarque-Bera test and the goodness-of-fit
test for normality is the same. This result is not surprising, as it is fairly well documented
that income distribution, in general, is skewed to the right (not normally distributed),
with a few households accounting for most of the total income. For this reason, we prefer
to use the median rather than the mean to get a more accurate reflection of the typical
income; hypothesis tests for the median will be discussed in Chapter 20.
EXERCISES 12.3
Concepts 22. The fo llo w in g freq u e n cy d is trib u tio n has a sample mean o f
- 3 . 5 and a sam ple standard de via tion o f 9.7.
21. Consider th e fo llo w in g sam ple data w ith m ean and standard
Class Frequency
de v ia tio n o f 20.5 and 5.4, respectively.
Less than − 10 70
Class Frequency
− 1 0 up to 0 40
Less th a n 10 25
0 up to 10 80
10 u p to 20 95
10 or more 10
20 up to 30 65
30 o r m o re 15 A t th e 1% significance level, use th e go o dn e ss-o f-fit te st fo r
n o rm a lity to d e te rm in e w h e th e r or n o t th e data are n o rm a lly
n = 200
distribu te d .
a. Using th e g o o d n e ss -o f-fit te st fo r n o rm a lity, specify the
23. You are given th e fo llo w in g sum m ary statistics fro m a sample
co m p e tin g hypotheses in order to d e te rm in e w h e th e r or
o f 50 observations:
n o t th e data are n o rm a lly d is trib u te d .
Mean 77.25
b. Calculate th e value o f th e te s t statistic.
Standard Deviation 11.36
c. At th e 5% significance level, w h a t is th e critical value?
Skewness 1.12
W h a t is th e decision rule?
Kurtosis 1.63
d. W h a t is th e conclusion?
a. Using th e Jarque-Bera test, specify th e n ull and Total Compensation
a lte rn a tive hypotheses to d e te rm in e w h e th e r o r n o t th e Observed Frequency
(in millions of $)
data are n o rm a lly d is trib u te d .
Less than 5 43
b. Calculate th e value o f th e te st statistic.
5 up to 10 65
c. A t th e 5% significance level, w h a t is th e critical
10 up to 15 32
value?
d. W h a t is th e conclusion? Can you co n clu de th a t 15 up to 20 38
Total = 300
Class (in percent) Observed Frequency
a. Using th e g o o d n e s s-o f-fit te st fo r no rm a lity, state Less than − 5 14
th e c o m p e tin g hypotheses in o rd e r to d e te rm in e if w e
− 5 up to 0 9
can reject th e professor's n o rm a lity claim.
0 up to 5 18
b. Calculate th e value o f th e te s t statistic.
5 up to 10 11
c. At a 5% significance level, w h a t is th e critical value?
d. W h a t is th e conclusion to th e test?
10 or more 8
n = 60
25. Fifty cities p ro vided in fo rm a tio n on vacancy rates (in percent)
in local a p a rtm e n ts in th e fo llo w in g freq u e n cy d istrib u tio n . Source: w w w .ya h oo .fina nce .com .
The sam ple m ean and th e sam ple standard de via tion are 9%
Over th is tim e period, th e fo llo w in g su m m a ry statistics are
and 3.6%, respectively.
provided:
Vacancy Rate (in percent) Frequency
9 up to 12 20
a. C o n d uct a g o o d n e ss -o f-fit te st fo r n o rm a lity at th e
12 or more 10 5% significance level. Can you co nclude th a t m o n th ly
returns do n o t fo llo w th e no rm a l d istribu tio n ?
A p p ly th e g o o d n e s s-o f-fit te st fo r n o rm a lity a t th e 5%
significance level. Do th e sam ple data suggest th a t vacancy b. C o n d uct th e Jarque-Bera test a t th e 5% significance
d is trib u tio n , a long w ith some s u m m a ry statistics. Total D e p o t in th e data file on th e te x t w ebsite, labeled
th e p o te n tia l value o f stock o p tio ns, and gains fro m stock a. Specify th e co m p e tin g hypotheses fo r th e Jarque-Bera
o p tio n s exercised. te st o f n o rm a lity o f Hom e Depot's stock prices.
b. Calculate th e value o f th e Jarque-Bera te s t statistic. Use a. Specify th e co m p e tin g hypotheses fo r th e Jarque-Bera
Excel to calculate th e p -value. te s t o f n o rm a lity o f MPG.
c. At α = 0.05 can you conclude th a t Hom e Depot's stock b. Calculate th e value o f th e Jarque-Bera te st statistic. Use
prices are n o t n o rm a lly distribu te d ? Excel to calculate th e p -value.
29. FILE (Use Excel) Access th e miles per g a llon (MPG) data on c. At α = 0.05 can you co nclude th a t MPG are n o t n o rm a lly
th e te x t w ebsite, labeled MPG. distribu te d ?
As part of a broader report concerning the mutual fund industry in general, three- Sample
year return data for the 50 largest mutual funds were collected with the objective
of determining whether or not the data follow a normal distribution. Information of Report—
this sort is particularly useful because much statistical inference is based on the as
sumption of normality. If the assumption of normality is not supported by the data, it Assessing
may be more appropriate to use nonparametric techniques to make valid inferences.
Table 12.A shows relevant summary statistics for three-year returns for the 50 largest Whether
mutual funds.
Data Follow
TABLE 1 2 .A Three-Year Return S u m m a ry M easures fo r th e 50 Largest M u tu a l Funds, A u g u s t 2008
n = Σo
=
i 50 Σp i = 1 n = Σ e i = 50
Table 12.C shows the competing hypotheses, the value of the test statistic, and the value
that result from applying the goodness-of-fit test for normality and the Jarque-Bera test
for normality.
Goodness-of-fit Test:
H0: Returns are norm ally distributed. χ22 = 10.53 P(χ22 ≥ 10.53) = 0.0052
HA: Returns are not norm ally distributed.
Jarque-Bera Test:
H0: S = O a n d tf = 0 χ22 = 29.62 P(χ22 ≥ 29.62) = 0.0000
HA: S = 0 or K = 0
At the 5% significance level, the p -value of 0.0052 from the goodness-of-fit test allows
us to reject the null hypothesis. The three-year returns do not follow the normal distri
bution at the 5% significance level.
Under the Jarque-Bera (J B ) test, the null hypothesis states that the skewness
coefficient and the kurtosis coefficient are both zero. The value for the JB test sta
tistic is 29.62 and its associated p -value is 0.0000; thus, at the 5% significance level,
the null hypothesis that skewness and kurtosis are both zero is rejected. This result
is consistent with the conclusion drawn from the goodness-of-fit test for normality.
Both statistical tests reject the null hypothesis of normality— three-year returns do not
follow the normal distribution. Statistical inference would best be conducted using
nonparametric techniques.
Conceptual Review
LO 12.1 Conduct a goodness-of-fit test for a m ultinom ial experim ent.
A multinomial experiment consists of a series of n independent trials of a random
experiment such that on each trial there are k possible outcomes or categories. We
conduct a goodness-of-fit test on a multinomial experiment to test if the population
proportions equal some predetermined (hypothesized) values.
For a multinomial experiment with k categories, the test statistic follows the
χ2df(chi-square) distribution with k − 1 degrees of freedom and its value is calculated
where oi and ei = n × pi are the observed frequency and expected
frequency, respectively, in the ith category. The test is valid when the expected fre
quencies in each category are five or more. This test is always implemented as a right
tailed test.
M arket Share Total Sales in 2011 (those born after 1984) were asked a b o u t th e ir preference for
Firm in 2010 (in billions of $) d elivering th e news a b o u t breaking up a relationship (The Wall
Street Journal, Novem ber 2 7 - 2 8 , 2010). One o f th e shocking
1 0.40 200
results was th a t o n ly 47% o f users preferred to break th e news
2 0.30 180
in person. A researcher decides to verify th e survey results o f
3 0.20 100 Facebook by ta king her o w n sample o f 200 young Facebook
4 0.10 70 users. The preference percentages fro m Facebook and the
Three nights Saturday night die if it happens afte r 11 pm . The stu d y investigated 58,593
One n ight Two nights
cardiac arrests d u rin g th e day or evening. O f those, 11,604
117 137 298 92
survived to leave th e hospital. There w ere 28,155 cardiac
a. Specify th e c o m p e tin g hypotheses to te st w h e th e r th e arrests d u rin g th e s h ift th a t began a t 11 p m , c o m m o n ly
recent p ro p o rtio n s d iffe r fro m those cited in th e study. referred to as th e graveyard shift. O f those, 4,139 survived fo r
b. Calculate th e value o f th e te s t statistic. discharge. The fo llo w in g c o n tin g e n c y ta b le sum m arizes the
32. A local TV sta tion claim s th a t 60% o f p e ople s u p p o rt Day or Evening Shift 11,604 46,989 58,593
C andidate A, 30% s u p p o rt C andidate B, and 10% s u p p o rt Graveyard Shift 4,139 24,016 28,155
C andidate C. A survey o f 500 registered voters is taken. The Column Totals 15,743 71,005 86,748
acco m p a n yin g ta b le indicates h o w th e y are likely to vote.
a. Specify th e c o m p e tin g hypotheses to d e te rm in e
Candidate A Candidate B Candidate C
w h e th e r a patient's survival depends on th e tim e at
350 125 25 w h ic h he/she experiences cardiac arrest.
and classifies th e stock prices in these industries in to one o f w h o h e a t w ith n a tu ra l gas o r e le c tric ity . A re p re se nta tive
St. John's w o rt 12 15
Over th is tim e period, th e fo llo w in g sum m a ry statistics are
Placebo 14 13
provided:
At th e 5% significance level, do th e data in dicate th a t
Mean Median Standard Deviation Skewness Kurtosis
St. John's w o r t affects ch ild re n w ith ADHD?
0.31% 0.43% 6.49% 0.15 0.38
37. A recent p o ll asked 3,228 Am ericans aged 16 to 21 w h e th e r
th e y are likely to serve in th e U.S. m ilita ry. The fo llo w in g
a. C on d uct a g o o dn e ss-o f-fit te s t fo r n o rm a lity a t th e 5%
table, cross-classified by g e n de r and race, reports those w h o
significance level. Can you conclude th a t m o n th ly stock
responded th a t th e y are likely or very likely to serve in th e
returns d o n o t fo llo w th e norm al d istribu tio n ?
a c tiv e -d u ty m ilitary.
b. C on d uct th e Jarque-Bera te st a t th e 5% significance
Race
level. Are y o u r results consistent w ith yo u r answer
Gender Hispanic Black W hite in p a rt a?
Male 1098 678 549
40. F IL E (Use Excel) Access th e data file on th e te x t w ebsite
Female 484 355 64 labeled A rlington Homes. It contains data on various
SOURCE: Defense Human Resources Activity telephone poll o f 3,228 Americans variables, in clu d in g price and square footage, fo r 36 sin g le
conducted October through December 2005.
fa m ily homes in A rlin g to n , Massachusetts, sold in th e first
a. State th e co m p e tin g hypotheses to te s t w h e th e r race q u a rte r o f 2009.
and g e n de r are d e p e n d e n t w h e n m aking a choice to a. Use th e Jarque-Bera te st to te st if house prices are n ot
serve in th e m ilitary. no rm a lly d is trib u te d a t α = 0.05.
b. C onduct th e te s t using th e critical value a pproach a t th e b. Use th e Jarque-Bera te s t to te st if square fo o tag e is no t
5% significance level. no rm a lly d is trib u te d a t α = 0.05.
CASE STUDIES
S o m e w h a t im p o r ta n t 0.25 140
N o t to o im p o r ta n t 0.15 96
D o n 't k n o w 0.02 4
0 up to 5 14
5 or more 18
He also calculates the following summary statistics over this time period:
LO 1 3.5 Conduct and evaluate hypothesis tests based on tw o-w ay ANOVA with
interaction.
In this chapter, w e study analysis o f variance w hich is more com m only referred to as ANOVA. ANOVA
is a statistical technique used to determ ine if differences exist between the means o f three or more
populations, thus generalizing the tw o-sam ple tests discussed in Chapter 10. For instance, we may
w a nt to determ ine w h ether all brands o f small hybrid cars have the same average miles per gallon.
Or we may wish to compare the effectiveness o f different fertilizers on the average yield per acre.
These are examples o f one-way ANOVA, where we examine the effect o f one categorical variable or
one factor on the mean. We then move on to tw o -w a y ANOVA, where the mean may be influenced
by tw o categorical variables or tw o factors. For instance, w e may w a nt to determ ine if average miles
per gallon are affected by the brand o f a hybrid and the octane level o f gasoline. Or w e may wish to
determine if the average yield per acre is influenced by the fertilizer and the acidity level o f the soil.
Tests based on tw o -w a y ANOVA can be conducted with or w ithout the interaction o f the factors.
I N T R O D U C T O R Y CAS E
Public Transportation
Sean C ox, a re s e a rc h a n a ly s t a t a n e n v ir o n m e n t a l o rg a n iz a tio n , b e lie v e s t h a t a n u p s w in g in t h e use o f p u b lic t r a n s p o r ta
t io n has ta k e n p la c e d u e t o e n v ir o n m e n t a l c o n c e rn s , t h e v o la t ilit y o f gas p rices, a n d t h e g e n e ra l e c o n o m ic c lim a te . H e
is e s p e c ia lly p le a s e d w it h a r e c e n t s tu d y , w h ic h h ig h lig h t s t h e a v e ra g e a n n u a l c o s t s a v in g s w h e n c o m m u te r s use p u b lic
t r a n s p o r ta t io n (The B o sto n G lobe, M a y 8 ,2 0 0 9 ) . C o m m u te r s w h o use p u b lic tr a n s p o r ta t io n save o n b u y in g , m a in t a in
in g , a n d o p e r a t in g t h e ir cars, w h ic h c o m p r is e t h e la rg e s t h o u s e h o ld e x p e n d it u r e a fte r h o u s in g . T h e s tu d y fin d s t h a t
B o s to n le a d s 2 0 o t h e r A m e ric a n c itie s in t h e a m o u n t t h a t c o m m u t e r s ca n save if t h e y ta k e p u b lic t r a n s p o r ta t io n . Sean
w o n d e r s w h e t h e r o r n o t c o s t s a v in g s v a ry d r a m a tic a lly b y city. H e c o lle c ts a r e p r e s e n ta tiv e s a m p le o f p u b lic tr a n s it
rid e rs in t h e t o p f o u r c o s t-s a v in g s c itie s : B o s to n , N e w York, San Francisco, a n d C h ic a g o . T a b le 13.1 s h o w s e a ch p u b lic
tr a n s it rid e r's a n n u a l c o s t s a v in g s b y c ity ; t h e d a ta a re a ls o a v a ila b le o n t h e t e x t w e b s ite , la b e le d P u b lic T r a n s p o r t a t io n .
Se a n w a n t s t o u s e t h e a b o v e s a m p le in f o r m a t io n to :
2. E s ta b lis h w h e r e m e a n c o s t s a v in g s s ig n if ic a n t l y d iffe r.
W e p r o v id e a s y n o p s is o f t h is c a s e a t t h e e n d o f S e c tio n 13.2.
15.1 One-Way ANOVA
LO 13.1 We use analysis of variance (ANOVA) tests to determine if differences exist between the
Provide a conceptual means of three or more populations, thus generalizing the two-sample tests discussed
overview of ANOVA. in Chapter 10. These tests are based on the F(df1,df2) distribution that was introduced in
Chapter 1 1. One-way ANOVA refers to comparing population means based on one cat
egorical variable or factor. For instance, in the introductory case of this chapter we want
to compare cost savings of using public transportation depending on where an individual
resides. We thus delineate cost savings of using public transportation by city (the factor).
We choose to collect sample data from four cities: Boston, New York, San Francisco,
and Chicago. In general, when performing a one-way ANOVA test, our sample consists
of c levels of the factor drawn from each of c populations. In our example c equals four.
In addition, we make the following three assumptions:
1. The populations are normally distributed.
2. The population standard deviations are unknown but assumed equal.
3. The samples are selected independently.
In short, each population i, where i = 1, 2, . . . , c, has mean μ i and standard deviation σ
where both parameters are unknown. From each population, we draw independent ran
dom samples where the selection of one sample does not affect the selection of another
sample. In other words, the experiment has a completely random ized design. From each
sample, we calculate the sample mean and the sample standard deviation si to imple-
ment the test.
The term “treatment” is often used to identify the c populations being examined. The
practice of referring to different populations as different treatments is due to the fact that
many ANOVA techniques were originally developed in connection with agricultural ex
periments where different fertilizers were regarded as different treatments applied to soil.
Table 13.2 summarizes the notation of the completely randomized design of ANOVA.
We define the notation as:
xij is the j th value in the ith sample,
is the sample mean of the ith sample,
si is the sample standard deviation of the ith sample, and
ni is the number of observations in the ith sample.
Sample Mean
Sample Standard Deviation S1 S2 ⋯ Sc
Sample Size n1 n2 ⋯ nc
It is easiest to examine one-way ANOVA through an example. Recall from the introduc LO 13.2
tory case that Sean wishes to determine whether the annual cost savings from using public C o n d u c t an d
transportation varies by city. He focuses on annual cost savings in Boston, New York, San e v a lu a te
Francisco, and Chicago. Table 13.3 reproduces the summary statistics from each city. h ypoth esis
T A B L E 1 3 . 3 S u m m a ry Statistics fro m Using Public T ra n s p o rta tio n in Four Cities tests based
o n o n e -w a y
Boston New York San Francisco Chicago
ANOVA.
Since we wish to test whether or not the mean annual cost savings from using public
transportation is the same in Boston, New York, San Francisco, and Chicago, we formu
late the following competing hypotheses:
H0: μ1 = μ2 = μ3 = μ 4
HA: Not all population means are equal
Note that HA does not require that all means must differ from one another. In principle, the
sample data may support the rejection of H0 in favor of HA even if only two means differ.
When conducting the equality of means test, you might be tempted to set up a series
of hypothesis tests, comparing μ1 and μ 2, then μ1 and μ 3, and so on, and then use the
two-sample t test discussed in Chapter 10. However, such an approach is not only cum
bersome, but also flawed. In this example, where we evaluate the equality of four means,
we would have to compare six combinations of two means at a time. This is based on the
combination formula first encountered in Chapter 4 that yields combina
tions. Also, by conducting numerous pairwise comparisons, we inflate the risk of the
Type I error α ; that is, we increase the risk of incorrectly rejecting the null hypothesis. In
other words, if we conduct all six pairwise tests at a 5% level of significance, the resulting
α for the overall test will be greater than 5%. Later, we will discuss how we can modify
the significance level of pairwise comparison tests for an accurate inference.
Fortunately, the ANOVA technique avoids this problem by providing one test that
simultaneously evaluates the equality of several means. In the public transportation
example, if the four population means are equal, we would expect the resulting sample
means, to be relatively close to one another. Figure 13.1a illustrates
the distributions of the sample means if H0 is true. Here, the relatively small variability
in the sample means can be explained by chance. What if the population means differ?
Figure 13.1b shows the distributions of the sample means if the sample data support HA.
In this scenario, the sample means are relatively far apart since each sample mean is
F IG U R E 1 3 .1 The lo g ic o f ANOVA
calculated from a population with a different mean. The resulting variability in the sample
means cannot be explained by chance alone.
We develop two independent estimates of the common population variance σ 2. One
estimate can be attributed to inherent differences between the c populations, while the
other estimate can be attributed to chance.
1. Between-Treatments Estimate of σ 2 . One estimate of σ2 is based on the variability
between the sample means. This is referred to as between-treatments variability.
2. Within-Treatments Estimate of σ 2 . The other estimate of σ2 is based on the
variability of the data within each sample, that is, the variability due to chance. This
estimate is generally called within-treatments variability.
If we find that the variability of between-treatments is significantly greater than the vari
ability within-treatments, then we are able to reject the null hypothesis of equal means;
this is equivalent to concluding that the ratio of between-treatments variability to within-
treatment variability is significantly greater than one.
Between-Treatments Estimate of σ 2
Between-treatments variability is based on a weighted sum of squared differences between
the sample means and the overall mean of the data set, referred to as the grand mean
and denoted as We compute the grand mean by summing all observations in the data
set and dividing by the total number of observations.
Each squared difference of a sample mean from the grand mean is multiplied
by the respective sample size for each treatment ni. After summing the weighted squared
differences, we arrive at a value called the sum of squares due to treatments or SSTR.
When we average a sum of squares over its respective degrees of freedom, we obtain the
mean sum of squares. Dividing SSTR by c − 1 yields the mean square for treatments or
MSTR.
C A L C U L A T I O N S F OR T H E B E T W E E N - T R E A T M E N T S E S T I M A T E O F σ 2
The calculations for SSTR, and MSTR for the public transportation example are as follows:
W ithin-Treatments Estimate of σ 2
We just calculated a value of MSTR equal to 4,401,573. Is this value of MSTR large
enough to indicate that the population means differ? To answer this question we compare
MSTR to the variability that we expect due to the nature of sampling. We first calculate the
e rro r sum of squares, denoted as SSE. SSE provides a measure of the degree of variabil
ity that exists even if all population means are the same. We calculate SSE as a weighted
sum of the sample variances of each treatment. We calculate the mean square error MSE
by dividing SSE by its respective degrees of freedom, d f = nT − c.
The values of SSE and M SE for the public transportation example are calculated as
follows:
The test statistic for the hypothesis test of the equality of the population means
using one-way ANOVA is assumed to follow the distribution with d f = c − 1
and df2 = nT − c and its sample value is
where MSTR and MSE are based on independent samples drawn from c normally
distributed populations with a common variance σ2 .
When conducting an ANOVA test, we reject the null hypothesis when the numerator of
the test statistic is significantly greater than the denominator. This result would
imply that between-treatments variability is significantly greater than within-treatments
variability, indicating that the population means are not all equal. This test is always a
right-tailed test. We are now in a position to conduct a four-step hypothesis test for the
public transportation example.
Given MSTR = 4,401,573, MSE = 7,209, d f 1 = c − 1 = 4 − 1 = 3, and df2 =
nT − c = 24 − 4 = 20, we compute the value of the test statistic as
Since the ANOVA test is a right-tailed test, the critical value with α = 0.05, d f 1= 3, and
d f 2= 20 is found from the F table as we show a portion of the
F table in Table 13.4 as a refresher. Hence, the decision rule is to reject H0 if F(3,20) > 3.10.
TA B LE 13.4 Portion of the F table
N um erator Degrees o f Freedom, d f 1
D e n o m in a to r D egrees A rea in
o f F re e d o m , d f2 U p p e r Tail 1 2 3
We reject the null hypothesis because the value of the test statistic falls in the rejection
region (610.57 is greater than 3.10). Therefore, we conclude that the mean cost savings
from using public transportation is not the same for each city.
Total SST nT − 1
We should also note that total sum of squares SST is equal to the sum of the squared dif
ferences of each observation from the grand mean. This is equivalent to summing SSTR
with SSE; that is, SST = SSTR + SSE.
ANOVA
Source o f V aria tio n SS df MS F p -v a lu e F Cr i t
Between Groups 13204720 3 4401573 610.57 7.96E-20 3.098
Within Groups 144180 20 7209
Total 13348900 23
EXAMPLE 13.1
Using the information in Table 13.6, repeat the ANOVA test for the public transpor
tation example using the p -value approach.
EXERCISES 1 3. 1
Concepts
1. A random sample of five observations from three treatments a. Calculate th e grand mean.
produced the following data: b. Calculate SSTR and MSTR.
T re a tm e n ts c. Calculate SSE and MSE.
A B C d. Specify th e c o m p e tin g hyp o th e se s in o rd e r to
22 20 19 d e te rm in e w h e th e r som e differences e xist b e tw e e n
25 25 22 th e p o p u la tio n means.
94 86 78
Total 117.72 79
89 89 79
a. Fill in th e m issing statistics in th e ANOVA table.
89 69 77
b. Specify th e c o m p e tin g hypotheses in o rd e r to d e te rm in e
83 79 78
w h e th e r som e differences exist b e tw e en th e p o p u la tio n
means.
c. At the 5% significance level, w h a t is the conclusion to the test?
Source o f
Variation SS df MS F p-value
Between a. Specify th e c o m p e tin g hypotheses in o rd e r to test the
We can improve upon the precision of this estimate by substituting MSE from the ANOVA
test for s2p. Recall that when we conduct an ANOVA test we assume that we are sampling
from populations that have the same population variance σ 2. We still apply the tdf distribu
tion, but we use the degrees of freedom corresponding to MSE, or d f = nT − c.
F I S H E R ' S C O N F I D E N C E I N T E R V A L F OR μ i − μj
Fisher’s 100(1 − α )% confidence interval for the difference between two popula
tion means μ i − μ j is given by
where the mean square error M SE is estimated from the ANOVA test.
EXAMPLE 13.2
Using the sample means and the ANOVA results from the public transportation
example of Section 13.1, calculate 95% confidence intervals for the difference be
tween all possible pairings of the four population means. Comment on the direction
and the significance of the differences at the 5% level.
In addition, MSE = 7,209 (derived from the ANOVA test in Section 13.1), nT − c =
24 − 4 = 20, and Table 13.7 shows the 95% confidence
intervals.
TABLE 13.7 Fisher's 95% C o n fid e n c e Inte rva ls fo r Exam ple 13.2
Population M ean Differences Confidence Interval
μ Boston N
−
μ ewYork
μ Boston − μSanFrancisco
μ Boston − μChicago
μ NewYork − μSanFrancisco
μ NewYork − μChicago
μ SanFrancisco − μChicago
The 95% confidence interval for μB o sto n − μN e w Y o rk is given by 37 ± 100.97,
which is [ − 63.97, 137.97]. Since this interval contains the value zero, we
cannot reject the null hypothesis, given by H0: μB oston − μN ew Y ork = 0, at the
5% significance level. In other words, the average cost savings from using
public transportation in Boston and in New York are not significantly
different.
For μ B o s t o n − μ san Francisco, the entire interval, ranging from 794.75 to 1,009.25, is
above the value zero. We can therefore conclude at the 5% significance level that av
erage cost savings from using public transportation in Boston are different from the
average cost savings in San Francisco. In fact, the remaining intervals are all above
zero, suggesting that average cost savings are different between the correspond
ing cities. In other words, we can conclude at the 5% significance level that aver
age cost savings from using public transportation are different between Boston and
Chicago, New York and San Francisco, New York and Chicago, and San Francisco
and Chicago.
These pairwise comparisons of the means use Fisher’s least significant difference (LSD)
method, although the LSD tests are motivated within a hypothesis-testing framework.
(Recall that we can conduct two-tailed tests with confidence intervals.) We can apply this
method only if the ANOVA test has rejected the null hypothesis of equal means. However,
as mentioned earlier, some problems are associated with inferring the equality of means
by conducting paired tests. In Example 13.2, we have six paired tests. Therefore, if we
use the 5% significance level for each test, the probability that we would make a Type I
error (incorrectly rejecting a null hypothesis of equal means) on at least one of these in
dividual tests will be greater than 5%. The more means we compare, the more the Type I
error becomes inflated.
One way to avoid this problem is to perform each individual paired test at a re
duced significance level, which ensures that the overall significance level for the
equality o f all means does not exceed α . The resulting confidence intervals are wider
and hence reduce the probability of incorrectly rejecting the null hypothesis of equal
means. However, this technique reduces the power o f the test and thus results in an
increased risk of a Type II error (incorrectly failing to reject a null hypothesis o f equal
means).
Tukey’s 100(1 − α )% confidence interval for the difference between two popula
tion means μ i − μ j is given by
The studentized range value varies with the significance level α , the number
of populations c, and nT − c. Table 13.8 shows a portion of the studentized range table;
Table 5 in Appendix A provides a more comprehensive table. For example, with α = 0.05,
c = 6, and nT − c = 19, we find q0.05,(6,19) = 4.47. With α = 0.01, c = 3, and nT − c = 20,
we find q0.01,(3,20) = 4.64. These values are in boldface in Table 13.8.
TABLE 13.8 P o rtio n o f Values fo r in Tukey's HSD M e th o d
c = n u m b e r o f m eans
nT −c α 2 3 4 5 6 7 8 9
0.05 2.96 3.59 3.98 4.25 4.47 4.65 4.79 4.92
19
0.01 4.05 4.67 5.05 5.33 5.55 5.73 5.89 6.02
0.05 2.95 3.58 3.96 4.23 4.45 4.62 4.77 4.90
20
0.01 4.02 4.64 5.02 5.29 5.51 5.69 5.84 5.97
When using Fisher’s LSD method at some stated significance level α , the probabil
ity of committing a Type I error increases as the number of pairwise comparisons
increases; that is, the likelihood of incorrectly rejecting the null hypothesis of equal
means in at least one of the pairwise comparisons is greater than α . By using the
studentized range distribution over the tdf distribution, Tukey’s HSD method ensures
that the probability of a Type I error equals a , irrespective of the number of pairwise
comparisons.
EXAMPLE 13.3
A consumer advocate in California is concerned with the price of a common generic
drug. Specifically, he feels that one region of the state has significantly higher prices
for the drug than two other regions. He divides the state into three regions and col
lects the generic drug’s price from 10 pharmacies in each region. He produces the
summary statistics and ANOVA results shown in Table 13.9
Total 267.37 29
a. At the 5% significance level, do differences exist between the mean drug prices
in the three regions?
b. If significant differences exist, use Tukey's HSD method to determine which
regions’ means differ at the 5% significance level.
S O L U T IO N :
a. In order to test differences between the mean drug prices in the three regions,
we specify the competing hypotheses as
The ANOVA table shows the value of the test statistic as F (2,27) = 21.17 with a
p -value of 2.95 × 10−6, or P(F2,27 ≥ 21.17) ≈ 0. Since the p -value is less than
the significance level of 0.05, we reject H0 and conclude that not all mean drug
prices are equal.
b. Since each sample size is the same (n1 = n2 = n3 = 10 and nT = 30), we can
use Tukey’s confidence intervals for balanced data. Given a = 0.05, c = 3, and
nT − c = 30 − 3 = 27, we find We use
to compute 95% confidence intervals for all pairwise differences
of the means. The results are shown in Table 13.10.
μ1 − μ2
μ1 − μ3
μ2 − μ3
The asterisk * shows that the confidence interval does not include the value
zero, thus indicating the corresponding means are different at the 5% signifi
cance level. The consumer advocate’s claim is supported by the data. At the
5% significance level, the average price of generic drugs in region 3 is differ
ent from the average prices in regions 1 and 2. At the 5% significance level,
the consumer advocate cannot conclude that average prices in regions 1 and 2
differ.
We also employed Tukey’s method for unbalanced data using the ANOVA results from
the public transportation example. While the resulting intervals (not reported) became
wider than those reported in Table 13.7, the inference regarding the population means
remains the same.
S Y N O P S I S OF I N T R O D U C T O R Y CASE
n1 = 5 n2 = 8 n3 = 6 n4 = 5
EXERCISES 13.2
Concepts 14. A c o m p le t e ly r a n d o m iz e d d e s ig n o f a n an a ly s is o f v a r ia n c e
e x p e r i m e n t p r o d u c e d a p o r t i o n o f t h e f o l lo w in g A N O V A
12. The following statistics are computed by sampling from
ta b le .
three normal populations whose variances are equal:
SUMMARY
a. Calculate 9 5 % co nfidence intervals for μ 1 − μ 2 , μ 1 − μ 3 , Groups Count Average
and μ 2 − μ 3 using Fisher's LSD approach. Column 1 6 0.57
b. Repeat the analysis with Tukey's HSD approach.
Column 2 6 1.38
c. Which of these tw o approaches would you use to
Column 3 6 2.33
determ ine w hether differences exist between the
population means? Explain.
ANOVA
13. The fo llo w in g statistics are calculated by sam p ling fro m four
Source o f
normal populations whose variances are equal:
Variation SS df MS F p-value F crit
m e a n s d iffe r a t α = 0 . 0 1 .
Total 14.46 17
b. U se T u key's H S D t e s t t o d e t e r m i n e w h ic h p o p u la t io n
m e a n s d if f e r a t α = 0 . 0 1 . a. C o n d u c t a n A N O V A t e s t a t t h e 5 % s ig n ific a n c e le v e l t o
c. Given your response to p a rt b, w h ich m eans significantly b. R epeat th e analysis w ith Tukey's HSD approach.
differ? c. W hich o f these tw o approaches is m ore reliable? Explain.
15. A c o m p lete ly rand om ized design o f an analysis o f variance d. Does th e data suggest th a t th e lean te ch n iq u e is
e x p e rim e n t produced a p o rtion o f th e fo llo w in g ANOVA im proving efficiency? Explain.
table.
17. The fo llo w in g Excel o u tp u t sum m arizes th e results for a
ANOVA Source o f
Between Groups 2997.11 3 999.04 15.54 1.2E-06 2.866 Within Groups 1302.41 50 26.05
Total 2336.92 52
Total 5311.82 39
a. A t th e 5 % significance level, can w e conclude th a t
a. Use Fisher's LSD test to d e te rm in e w h ich m eans differ at average m p g differs b e tw e e n th e hybrids?
b. Use Tukey's HSD test to d e te rm in e w hich m eans differ at 5 % significance level to d e te rm in e w hich hybrids' m eans
c. Given yo ur responses to parts a an d b, d o th e p o pu lation 18. Do en erg y bills vary dram atically d e p e n d in g on w h e re you
m eans differ at th e 5% significance level? live in th e U nited States? Suppose 25 households from four
regions in th e U nited States are sam pled. Th e values for
A p p lic a tio n s th e average annual en ergy bill are show n b e lo w and are
16. In an a t te m p t to im p ro ve efficiency, Starbucks has consistent w ith those fo u n d by The D e p a rtm e n t o f Energy
11 ,000 U.S. stores (The W all Street J o u rn a l, A ugust 4 , 20 0 9 ). By Region West Northeast M idwest South
reducing th e tim e baristas (em ployees) spend on bending, Average Annual $1,491 $2,319 $1,768 $1,758
reaching, an d w alking , th e y will have m ore tim e to interact Energy Bill
w ith customers and im prove th e Starbucks experience.
Suppose Starbucks ad op ts th e lean te ch n iq u e at Store 1, b u t A p o rtion o f th e ANOVA calculations are below:
Source o f Variation F
makes no changes a t Stores 2 an d 3. On a recent M o n d a y SS df MS p-value
m o rnin g b e tw e e n th e hours o f 7:00 AM an d 8:00 AM, th e Between Groups 7531769 3 ? ? 7.13E-24
fo llow ing statistics w e re o b ta in e d relating to average tim e ?
Within Groups 3492385 96
per order (in seconds):
Total 11024154 99
a. C o m p le te th e ANOVA table.
Two-way A N O VA tests are used to simultaneously examine the effect of two factors
on the mean. These tests can be conducted with or without the interaction of the factors.
In the following example we initially apply one-way ANOVA and recognize its limita
tions. We then introduce two-way ANOVA without interaction. In Section 13.4 we dis
cuss two-way ANOVA with interaction.
EXAMPLE 13.4
Julia Hayes is an undergraduate who is completely undecided as to what career she
should pursue. To help in her decision process, she wants to determine whether or
not there are significant differences in annual incomes depending on the field of em
ployment. Initially, she confines her analysis to the following three fields: educational
services, financial services, and medical services. As a preliminary experiment, she
surveys four workers from each of these three fields and asks how much he/she earns
annually. Table 13.12 shows the results (in $ 1,000s) from the experiment; the data
are also available on the text website, labeled One-Factor Income.
Use one-way ANOVA to determine if differences exist among the fields’ average
incomes at the 5% significance level.
ANOVA
Source o f Variation SS df MS F p-value F crit
Between Groups 579.5 2 289.75 0.32998 0.727281 4.2565
Total 8482.25 11
The value of the test statistic is F(2,9) = 0.33. Using the critical value approach, the
decision rule is to reject H0 if F(2,9) > 4.26. We do not reject the null hypothesis since
the value of the test statistic does not fall in the rejection region (0.33 is not greater
than 4.26). At the 5% significance level, we cannot conclude that average incomes
differ by field.
Julia is surprised by these results, since she feels that those in the educational
services industry probably earn less than those in the other two fields. Julia is
advised that she must interpret these results with caution because many other fac
tors influence annual income— one of which is an individual’s educational attain
ment. We can capture the true influence of field of employment on income only
when educational attainment is held fixed.
As mentioned earlier, two-way ANOVA helps us find a more discriminating F(df1,df2) test
statistic, since the additional factor reduces the resulting error sum of squares. An added
requirement for two-way ANOVA tests is that all groups must have the same sample
size.
To show how two-way ANOVA works, we redo the analysis from Example 13.4,
but this time we allow the variation in income to be affected by field (the treatment or
factor A) and educational attainment (the block or factor B). We match a worker from
each field according to his or her highest educational attainment; that is, we delineate one
worker whose highest educational attainment is a high school degree; we then select three
other matched groups, or blocks, whose highest educational attainments are a bachelor’s
degree, a master’s degree, and a Ph.D. or its equivalent. The outcomes in this experiment
are matched or blocked according to their educational attainment, with one worker for
each field of employment and level of educational attainment. In general, blocks are the
levels at which we hold an extraneous factor fixed, so that we can measure its contribution
to the total variation of the data. This experimental design is called a randomized block
design. The experiment in this example is designed to eliminate the variability in income
attributable to differences in educational attainment.
TW O -W A Y A N O V A (W IT H O U T IN TERACTIO N)
AND R A N D O M IZE D BLOCK DESIGN
In two-way ANOVA without interaction, the total variability of the data is partitioned
into three components attributed to the treatment (factor A), the block (factor B),
and chance, respectively. The term block refers to a matched group of observations
from each population. If units within each block are randomly assigned to each of
the treatments, then the design of the experiment is referred to as a randomized
block design.
Table 13.14 shows the incomes (in $l,000s) for 12 workers according to their field of
employment and highest educational attainment; these data labeled Two-Factor Income
are also available on the text website. Also included in the table are the factor means.
High School 18 25 26
Bachelor's 35 45 43
Master's 46 58 62
Ph.D. 75 90 110
Factor A Means
Dividing SSA by the number of columns in the randomized block design minus one (c − 1)
yields the mean square for factor A , MSA. For this example, c equals three, so MSA is
Dividing SSB by the number of rows in the randomized block design minus one (r − 1)
yields the mean square for factor B , MSB, or
We can make some generalizations about the difference in the magnitudes of the SSE val
ues for the one-way ANOVA versus the two-way ANOVA examples. When we used one
factor (field of employment) to explain annual incomes, the value of SSE was 7,902.75
(see Table 13.13). By ignoring the second factor (level of educational attainment), we
could not establish that annual incomes were different by field of employment. However,
once we include this second factor, the value of SSE declines dramatically to 269.11. We
will show shortly that by accounting for the effect of educational attainment on income,
an F(df1,df2) test statistic is produced that allows Julia to conclude that significant differences
do exist among annual incomes by field of employment.
Dividing SSE by its degrees of freedom (nT − c − r + 1) yields the mean square
error M SE or
Most software packages easily provide these statistics. Table 13.15 shows the general
format of an ANOVA table for the randomized block design.
TA B LE 1 3 .1 5 General Format of ANOVA Table for Randomized Block Design
Source o f
Variation SS df MS F
Rows SSB r −1
Columns SSA c −1
Total SST nT − 1
Table 13.15 shows values for two F(df1,df2) test statistics. We use the first statis
tic to test whether sig
nificant differences exist between the factor B means. We use the second statistic
to test whether significant
differences exist between the factor A means.
Table 13.16 shows a portion of the results. Any differences between our manual calcula
tions and the values that appear in the table are due to rounding. Note that Excel provides
precise p -values for the calculated test statistics and critical values at the 5% significance
level (this is the value of α specified in Excel’s dialog box).
TABLE 13.16 Excel's ANOVA O utp ut for the Two-Factor Income Example
ANOVA
Source o f
Variation SS df MS F p-value F crit
Rows 7632.92 3 2544.31 56.58 8.6E-05 4.76
Total 8482.25 11
EXAMPLE 13.5
Use the Excel output from Table 13.16 to conduct the following hypothesis tests.
a. At the 5% significance level, do average annual incomes differ by field of
employment?
b. At the 5% significance level, do average annual incomes differ by level of edu
cational attainment?
S O L U T IO N :
a. Using the critical value approach, we determine whether average annual in
comes differ by field of employment. The competing hypotheses are
When testing whether the factor A (column) means differ, the value of the test
statistic is where d f1 = c − 1 = 3 − 1 = 2 and
df2 = nT − c − r + l = 12 − 3 − 4 + 1 = 6, that is, F(2,6) = 6.44. At the 5%
significance level, the critical value is F0.05,(2,6) = 5.14. The decision rule is to
reject H0 if F(2,6) > 5.14. We reject H0 because the value of the test statistic is
greater than the critical value. Therefore, contrary to the results derived earlier
with one-way ANOVA, average annual salaries do differ by field of employ
ment at the 5% significance level.
b. Using the p -value approach, we determine whether average annual incomes
differ by level of educational attainment. The competing hypotheses are
When testing whether the factor B (row) means differ, the value of the test
statistic is where d f1 = r − 1 = 4 − 1 = 3
and df1 = nT − c − r + 1 = 12 − 3 − 4 + 1 = 6, that is, F(3,6) = 56.58. From
Table 13.16, the p -value associated with this test statistic is 8.6 × 10−5, or
P(F(3,6) ≥ 56.58) ≈ 0.0000. We reject H0 because the p -value is less than the
significance level of 5%. We conclude at the 5% significance level that aver
age annual salaries also differ by level of educational attainment. Since level
of education attainment exerts a significant influence on salaries, it must be
incorporated in ANOVA testing.
We would like to point out that, analogous to the last section, we can apply the MSE
estimate from the two-way ANOVA test to construct useful confidence intervals for the
paired differences in population means using Fisher’s LSD method. The only significant
modification to these confidence intervals is with respect to degrees of freedom, which are
now given by d f = nT − c − r + 1. Similarly, we can also use Tukey’s HSD method to de
termine which column means or row means are significantly different from one another.
The value for the margin of error in the confidence interval will depend on whether we are
assessing differences between the column means or the row means. When constructing
the confidence interval for the difference between two column means, we calculate the
margin of error as where n is the number of observations in each column.
When constructing the confidence interval for the difference between two row means, we
calculate the margin of error as where n is the number of observations in
each row.
EXERCISES 13.3
Concepts c. A t t h e 5 % s ig n ific a n c e le v e l, c a n y o u c o n c lu d e t h a t t h e
c o lu m n m e a n s d iffer?
2 1 . T h e o b s e r v a t io n s s h o w n in t h e a c c o m p a n y in g t a b le w e r e
2 4 . For a r a n d o m iz e d b lo c k d e s ig n w i t h n o in t e r a c t io n h a v in g
d r a w n f r o m a r a n d o m iz e d b lo c k d e s ig n w i t h n o in t e r a c t io n .
f o u r le v e ls o f fa c t o r A a n d t h r e e le v e ls o f f a c t o r B, t h e results
Factor A
1 in c lu d e t h e f o llo w in g s u m o f s q u a r e te rm s :
Factor B 2 3 4
1 2 3 2 4 2.8
a. Construct an ANOVA ta ble.
2 6 5 7 6 6.0
b. A t th e 1% significance level, can you co nclude th a t th e
3 8 10 9 10 9.3
5.3 facto r A m eans differ?
6 .0 6.0 6.7
c. At th e 1% significance level, can you co nclude th a t th e
a. Calculate SST, SSA, SSB, an d SSE. facto r B m eans differ?
b. Calculate MSA, MSB, an d MSE. 25. Th e fo llo w in g ta b le sum m arizes a p o rtion o f th e results for a
c. Construct an ANOVA table. ran d o m ized block e x p e rim e n t w ith no interaction.
2 2 10 8 6.7 11
Total 186.92
3 0 −9 −2 − 3.7
a. Find th e missing values in th e ANOVA table.
4 −3 − 14 −8 − 8.3
b. At th e 5 % significance level, can you conclude th a t th e
1.0 0.5 2.5
co lum n m eans differ?
a. Calculate SST, SSA, SSB, a n d SSE. c. A t th e 5 % significance level, can you conclude th a t th e
b. Calculate MSA, MSB, an d MSE. row m eans differ?
c. Construct an ANOVA table. 2 6 . T h e f o l lo w in g t a b le s u m m a r iz e s a p o r t io n o f t h e results f o r a
d. A t a 1% significance level, can you conclude th a t th e ran d o m ized block e x p e rim e n t w ith no interaction.
colum n m eans differ?
ANOVA
e. At a 1% significance level, can you conclude th a t th e row
Source o f
m eans d iffer?T hat is, is blocking necessary?
Variation SS df MS F p-value F crit
2 3 . For a ra n d o m iz e d blo c k d e s ig n h a v in g t h r e e tr e a tm e n ts a n d fiv e
Rows 1057 5 MSB= ? FFactor B= ? 0.0064 3.326
blocks, t h e results in c lu d e t h e f o llo w in g s u m o f s quare term s:
Columns 7 2 MSA= 1 FFactor A= ? 0.9004 4.103
SST = 31 1 .7 SSA = 2 0 1 .6 SSE = 69.3
Error 330 10 MSE= 1
a. Construct an ANOVA table.
a r a n d o m iz e d b lo c k e x p e r i m e n t w i t h n o in t e r a c t io n in w h ic h Restaurant
fa c to r A consists o f fo u r d if f e r e n t k in d s o f o r g a n ic fe rtiliz e rs , Inspector 1 2 3
fa c to r B consists o f t h r e e d i ff e r e n t k in d s o f soil a c id ity levels, 1 72 54 84
a n d t h e v a r ia b le m e a s u r e d is t h e h e ig h t (in in c h e s ) o f a p la n t
2 68 55 85
at th e e n d o f fo u r w eeks.
3 73 59 80
ANOVA
4 69 60 82
Source of
5 75 56 84
Variation SS df MS F p-val ue F crit
Rows 0.13 2 MSB = ? FFactor B =
? 0.8182 5.143 a. U se Excel t o g e n e r a t e t h e A N O V A t a b le .
a v e r a g e g ra d e s d iffe r b y re s ta u ra n t?
Error 1.88 6 MSE = ?
c. If a v e r a g e g ra d e s d iffe r b y r e s ta u r a n t, use T u key's H S D
t e s t a t t h e 5 % s ig n ific a n c e le v e l t o d e t e r m in e w h ic h
Total 46.25 11
a v e r a g e s d iffer.
a. Find th e missing values in th e ANOVA table. d. A t t h e 5 % s ig n ific a n c e le v e l, c an y o u c o n c lu d e t h a t t h e
b. A t th e 5 % significance level, can you conclude th a t a v e r a g e g r a d e s d iffe r b y in s p e c to r? D o e s t h e m a y o r h a v e
average g ro w th o f th e p lan t differs by organic fertilizer? c a u se f o r c o n c e rn ?
U.S. O p e n . b a n k a s k e d t h e t h r e e a p p ra is e rs t o v a lu e (in $ 1 ,0 0 0 s ) th r e e
Round d i f f e r e n t ty p e s o f h o m e s : a c a p e , a c o lo n ia l, a n d a ra n c h . T h e
a. U se Excel t o g e n e r a t e t h e a p p r o p r ia t e A N O V A ta b le .
The fo llo w in g statistics w ere co m p uted:
b. A t t h e 5 % s ig n ific a n c e le v e l, c a n y o u c o n c lu d e t h a t
SST = 2 7 2.95 SSB = 93.2 SSE = 127.6 a v e r a g e v a lu e s d iffe r b y a p p ra is e r? S h o u ld t h e b a n k b e
a. Construct an ANOVA table. c o n c e rn e d w i t h a p p r a is e r in co ns is te nc ie s?
While we still use a randomized block design, we need at least two observations for each
combination of the ith level of factor A and the j th level of factor B. In other words, we
need more than one observation per cell.
To illustrate two-way ANOVA with interaction, we reanalyze the income example,
using new data with three incomes for each combination. Given the data in Table 13.17,
we ultimately want to determine whether interaction is present between educational at
tainment and field of employment; these data, labeled Two-Factor Income Interaction, are
also available on the text website.
20 27 26
High School 25 25 24
22 25 25
30 44 42
Bachelor's 35 46 43
34 48 45
46 50 62
Master's 47 58 56
50 56 60
79 90 90
Ph.D. 78 92 100
74 95 105
We are specifically interested in whether field of employment and education level inter
act with respect to average annual income. In order to find the relevant sum of squares
for the test, we first compute the cell means and the factor means. For example, the cell
mean for workers in educational services with a high school education is computed as
(20 + 25 + 22)/3 = 22.33. Factor means are based on one row or one column of the
data. Table 13.18 shows the cell means factor means and the grand mean
for the data.
T A B L E 1 3 .1 8 C ell a n d F a c to r M e a n s f o r T w o -F a c to r In c o m e E x a m p le w i t h In te r a c tio n
The Sum of Squares for Factor A, SSA, and the Sum of Squares for
Factor B, SSB
The calculations of SSA and SSB are analogous to the earlier two-way ANOVA dis
cussion, with one minor modification. For two-way ANOVA without interaction, SSA
and SSB were calculated as respectively. Now each for
mula is multiplied by the number of observations per cell w: and
Given the means in Table 13.18 with w = 3, c = 3, and r = 4, we
calculate
and
We divide by the respective degrees of freedom to obtain the mean square for factor A,
MSA and the mean square for factor B, MSB as
The Sum of Squares for the Interaction of Factor A and Factor B, SSAB
When two factors interact, the effect of one factor on the mean depends upon the specific
value or level present for the other factor. Interaction exists between these factors when
two mathematical expressions, denoted Expression 1 and Expression 2, are significantly
different from one another.
Expression 1 is defined as the difference of a cell mean from the grand mean, or
equivalently, Using the data from Table 13.18, one such difference would be
The sum of squares for the interaction between factor A and factor B, SSAB, is then based
on a weighted sum of the squared interactions (I 2) where the weight equals the number
of observations per cell w:
We obtain the mean square for interaction, MSAB, by dividing SSAB by the relevant
degrees of freedom (c − 1)(r − 1), or
Finally, we divide SSE by its degrees of freedom rc(w − 1) and obtain the mean square
error M SE as
TABLE 13.19 Excel's ANOVA O utp ut for the Two-Factor Income (with Interaction) Example
ANOVA
Source o f
V a ria tio n SS df MS F p -v a lue F cr i t
Total 22008 35
Table 13.19 shows values for three F(df1,df2) test statistics. The first two test statistics are
used to examine the main effects— potential differences in factor B or the row means
and potential differences in factor A
or the column means The third
calculated test statistic is
used to test whether there is interaction between factor A and factor B.
EXAMPLE 13.6
Use the Excel output from Table 13.19 to determine whether field of employment
and education level interact with respect to the average annual income at the 5%
significance level.
Note that due to the interaction, the differences between educational attainment are not the
same for all fields of employment. Such an outcome serves to complicate the interpreta
tion of the main effects, since differences in one factor are not consistent across the other
factor. This is why we should perform the interaction test before making any conclusions
using the other two F(df1,df2) statistics. If the interaction effect is not significant, then we
can proceed by focusing on the main effects: testing whether or not the row means or the
column means differ. If the interaction effect is significant, as it is here, one option is to
use another technique called regression analysis. Regression analysis is discussed in the
next four chapters.
EXERCISES 13.4
Concepts ANOVA
t h a t t h e r e is in t e r a c t io n b e t w e e n f a c t o r A a n d
Total 252.393 23
f a c t o r B?
c. A t t h e 1 % s ig n ific a n c e le v e l, c an y o u c o n c lu d e t h a t t h e
a. A t th e 1 % significance level, can you conclude th a t th ere
f a c t o r A m e a n s differ?
is interaction b e tw e e n th e t w o factors?
d. A t t h e 1 % s ig n ific a n c e le v e l, c an y o u c o n c lu d e t h a t t h e
b. Are you ab le to co n d u ct tests based on th e m ain effects?
t h r e e le v e ls o f f a c t o r B. Five r e p lic a te s f o r e a c h c o m b in a t i o n
A p p lic a tio n s
a re e x a m in e d . T h e results in c lu d e t h e f o l lo w in g s u m o f
35. Th e effects o f d e te rg e n t brand n a m e (factor A ) and th e
s q u a r e te rm s :
te m p e ra tu re o f th e w a te r (factor B) on th e brightness of
w as hed fabrics are being studied. Four brand nam es and
a. C onstruct an ANOVA ta b le
tw o te m p e ra tu re levels are used, an d six replicates for each
b. A t th e 5 % significance level, can you conclude th a t th ere co m b in a tio n are exam in ed . Th e fo llo w in g ANOVA ta b le
is interaction b e tw e e n facto r A an d facto r B? is produced.
c. A t th e 5 % significance level, can you conclude th a t th e
ANOVA
fa c to rm
A eans differ?
Source o f
d. A t th e 5 % significance level, can you conclude th a t th e
Variation SS df MS F p-value F crit
fa cto r B m eans differ?
Sample 75 1 75 63.38 8.92E-10 4.084
33. A researcher conducts a ran d o m iz ed block e x p e rim e n t
Columns 130.25 3 43.42 36.69 1.45E-11 2.839
w ith in teraction and provides th e fo llo w in g ANOVA
Interaction 8.67 3 2.89 2.44 0.0783 2.839
table.
Within 47.33 40 1.18
ANOVA
Source o f Total 261.25 47
Variation SS df MS F p-value F crit
a. C a n y o u c o n c lu d e t h a t t h e r e is in t e r a c t io n b e t w e e n
Sample 752.78 2 MSB= ? FFactor B= ? 0.0116 3.885
d e te rg e n t b ran d n a m e a n d th e te m p e ra tu re o f th e w a te r
Columns 12012.50 1 MSA= ? FFactor A =
? 5.62E-09 4.747
a t t h e 5 % s ig n ific a n c e level?
Interaction 58.33 2 M S A B = ? FInteraction ?
= 0.6117 3.885
b. A re y o u a b le t o c o n d u c t te sts b a s e d o n t h e m a in e ffects?
Within 683.33 12 MSE= ?
If yes, c o n d u c t th e s e te s ts a t t h e 5 % s ig n ific a n c e le v e l. If
no , e x p la in .
Total 1 3 5 0 6 .9 4 17 3 6 . F I L E A c o n s u m e r a d v o c a te e x a m in e s w h e t h e r t h e lo n g e v it y
o f c a r b a t t e r ie s (m e a s u r e d in y ea rs ) is a f f e c t e d b y t h e b r a n d
a. Find t h e m is s in g v a lu e s in t h e A N O V A ta b le .
n a m e ( fa c to r A ) a n d w h e t h e r o r n o t t h e c a r is k e p t in a
b. A t t h e 5 % s ig n ific a n c e le v e l, c a n y o u c o n c lu d e t h a t t h e r e g a r a g e ( fa c to r B ) . In t e r a c t io n is s u s p e c te d . T h e results a re
is a n in t e r a c t io n e ffec t? s h o w n in t h e a c c o m p a n y in g t a b l e a n d a re also a v a ila b le o n
c. A t t h e 5 % s ig n ific a n c e le v e l, c a n y o u c o n c lu d e t h a t t h e t h e t e x t w e b s it e , la b e le d B r a n d a n d G a r a g e .
c o lu m n m e a n s d iffer?
Brand Name of Battery
d. A t t h e 5 % s ig n ific a n c e le v e l, c an y o u c o n c lu d e t h a t t h e A
Kept in Garage? B C
r o w (s a m p le ) m e a n s differ?
Yes 7 ,8 ,8 6 ,7 ,7 8 ,9 ,9
3 4 . A r e s e a rc h e r c o n d u c t s a r a n d o m iz e d b lo c k e x p e r i m e n t
No 5 ,6 ,6 4 ,5 ,4 6 ,7 ,7
w i t h in t e r a c t io n a n d p r o v id e s t h e f o l lo w in g A N O V A
< 3 .0 66 48 42
58 59 52
66 65 50
60 64 58 56 56
⋮ ⋮ ⋮ ⋮ ⋮
68 57 57 59 56
Sample Does traffic congestion vary by city? The 2009 Annual Urban Mobility Report found
that traffic congestion, measured by annual hours of delay per traveler, was the worst
Report— in Los Angeles, followed by Washington, DC, Atlanta, Houston, and then San Fran
cisco. An independent survey was conducted to verify some of the findings. Twenty-
Evaluating five travelers in each of these cities were asked how many hours they wasted in traffic
over the past calendar year. Table 13.A reports the summary statistics. The sample data
Traffic indicate that Los Angeles residents waste the most time sitting in traffic with an aver
age of 69.2 hours per year. Washington, DC, residents rank a close second, spending an
Congestion average of 62 hours per year in traffic. Residents in Atlanta, Houston, and San Francisco
spend on average, 57.0, 56.5, and 55.6 hours per year in traffic. Houston had the highest
by City and San Francisco the lowest variability of congestion as measured by their respective
standard deviations.
n 1 = 25 n 2 — 25 n 3 = 25 n 4 = 25 n 5 = 25
LO 13.3 Use confidence intervals and Tukey's HSD m ethod in order to determ ine which
means differ.
The ANOVA test can determine whether significant differences exist between the popu
lation means. However, it cannot indicate which population means differ. By construct
ing confidence intervals for all pairwise differences of the population means, we can
identify which means differ. Note that if the ANOVA test does not reject H0, then there
are no differences to find.
Fisher’s LSD method is implemented by computing 100(1 − α )% confidence
intervals for all mean differences where
the mean square error MSE is estimated from the ANOVA test. If the computed interval
does not include the value zero, then we reject the null hypothesis H0: μ i − μ j = 0.
When pairwise comparisons are made with Fisher’s LSD method, we inflate the risk of
the Type I error a ; that is, we increase the risk of incorrectly rejecting the null hypoth
esis. In other words, if we conduct all pairwise tests at a = 0.05, the resulting a for the
overall test will be greater than 0.05.
A more powerful multiple comparison technique is Tukey’s HSD method, which
seeks out “honestly significant differences” between paired means. Tukey’s method
uses the studentized range distribution, which has broader, flatter, and thicker tails
than the tdf distribution, and therefore, protects against an inflated risk of a Type I
error. Tukey’s 100(1 — a)% confidence interval for μ i − μ j is computed as
for balanced data (n = ni = nj) and
for unbalanced data (ni ≠ nj), where qα,(c,nT–c) is the studentized range value.
LO 13.4 Conduct and evaluate hypothesis tests based on two-way ANOVA with no interaction.
Whereas one-way ANOVA tests are used to compare population means based on one
factor, two-way ANOVA tests extend the analysis to measure the effects of two fac
tors simultaneously. The additional factor explains some of the unexplained variation, or
equivalently, reduces the error sum of squares for a more discriminating F(df1,df2) statistic.
Two-way ANOVA tests can be conducted with or without interaction between the factors.
In two-way ANOVA without interaction, we partition the total variation SST into the
sum of squares for factor A (SSA), the sum of squares for factor B (SSB), and the error
sum of squares (SSE), that is, SST = SSA + SSB + SSE. We find two statistics. The
value of the first statistic where d f 1= r – 1 and df2 – nT – c – r + 1
is used to test whether significant differences exist between the factor B means (the
row means). The value of the second statistic where d f 1 = c – 1 and
df2 = nT – c – r + 1 is used to test whether significant differences exist between the
factor A means (the column means).
LO 13.5 Conduct and evaluate hypothesis tests based on tw o -w ay ANOVA w ith interaction.
In two-way ANOVA with interaction, we partition the total variation SST into
four components: the sum of squares for factor A (SSA), the sum of squares
for factor B (SSB), the sum of squares for the interaction of the two factors
(SSAB), and the error sum of squares (SSE), that is, SST = SSA + SSB + SSAB
+ SSE. Here we find the values of three F(df1,df2) test statistics. The first two sta
tistics are used to examine the m ain effects— differences in the levels of fac
tor where d f 1 = r – 1 and d f 2 = rc(w – 1)) and differences in
the levels of factor A where d f 1 = c – 1 and df2 = rc(w – 1)).
The value of the third test statistic where d f 1 = (r – 1)(c – 1)
and df2 = rc(w – 1), is used to test whether there is interaction between factor A and
factor B.
Interaction between the factors complicates the interpretation of the main effects. This
is why we should perform the interaction test before testing the main effects. If the in
teraction effect is not significant, then we can proceed by focusing on the main effects.
If the interaction effect is significant, one option is to use another technique called
regression analysis. Regression analysis is discussed in the next four chapters.
o p e r a t o r s d iffe r. A r a n d o m s a m p le o f f iv e e m p lo y e e s in e a c h d iffer?
o f t h e f o u r c a te g o rie s y ie ld s t h e s a la ry d a t a g iv e n in t h e t a b le 4 0 . A n e c o n o m is t w a n t s t o d e t e r m i n e w h e t h e r a v e r a g e P r ic e /
b e lo w : E a rn in g s (P /E ) ratio s d if f e r f o r fir m s in t h r e e in d u s trie s .
Average Salaries o f Transportation Operators ($1,0 0 0 s) In d e p e n d e n t s a m p le s o f f iv e fir m s in e a c h in d u s t r y s h o w t h e
le v e l t o d e t e r m in e w h ic h in d u s t r ie s 'm e a n P /E ratios
a. C o n s tr u c t a n A N O V A t a b l e a n d e s t i m a t e t h e p - v a lu e . differ.
o p e r a t o r s d iffer. o n e o f th r e e routes a n d w o n d e r s w h e t h e r t h e a v e ra g e c o m m u t e
tim e s (in m in u te s ) d iffe r b e t w e e n t h e t h r e e routes. H e o b tains o u tp u t fo r α = 0 .0 5 s u m m a r iz e s a p o r t io n o f t h e results f o r a
t h e fo llo w in g d a ta a fte r tra v e lin g e ac h ro u te fo r o n e w e e k . c o m p le t e ly r a n d o m iz e d d e s ig n in w h ic h t h e t r e a t m e n t s w e r e
29
R ou te 1 30 33 30 32 t h r e e d i f f e r e n t s e c tio n s o f a n a c c o u n t in g c o u rs e a n d t h e
27 v a r ia b le m e a s u r e d w a s t h e g r a d e o n t h e fin a l e x a m .
R o u te 2 32 28 30 29
25
R ou te 3 27 24 29 26
ANOVA
T h e f o ll o w in g o n e - w a y A N O V A resu lts w e r e o b t a i n e d f o r α = 0 .0 1 : Source o f
ANOVA Variation SS df MS F p-value F crit
b. A t t h e 5 % s ig n ific a n c e le v e l, c a n y o u c o n c lu d e
a. D e t e r m i n e a t t h e 1 % s ig n ific a n c e le v e l w h e t h e r t h e
t h a t a v e r a g e g r a d e s d iffe r in t h e a c c o u n t in g sectio ns?
a v e r a g e c o m m u t e t im e s d iffe r b e t w e e n t h e t h r e e ro u te s .
tim e s d iffer. w i t h o u t i n t e r a c t i o n w h e r e f a c t o r A ( c o lu m n ) r e p r e s e n t s
t h r e e i n c o m e c a t e g o r i e s (lo w , m e d i u m , h i g h ) , f a c t o r B
4 2 . B e fo re t h e rec e s sion, jo b - c r e a t in g cities in t h e S u n b e lt, like
( r o w s ) c o n s is ts o f t h r e e d i f f e r e n t k in d s o f p o lit i c a l p a r t ie s
Las V e g a s , P h o e n ix , a n d O r la n d o s a w t h e ir p o p u la tio n s ,
( D e m o c r a t , R e p u b l ic a n , I n d e p e n d e n t ) , a n d t h e
in c o m e levels, a n d h o u s in g pric e s s u rg e . Las V e ga s , h o w e v e r ,
v a r ia b le m e a s u r e d w a s t h e a m o u n t (in $) c o n t r i b u t e d
o f f e r e d s o m e t h in g t h a t o f t e n e lu d e d t h e s e o t h e r cities:
t o t h e p o lit ic a l p a r t y d u r i n g t h e 2 0 0 8 p r e s i d e n t ia l
u p w a r d m o b i lit y f o r t h e w o r k in g class. Fo r e x a m p l e , h a r d
e le c t i o n .
w o r k in g h o t e l m a id s w e r e a b le t o p r o s p e r d u r in g t h e b o o m
P hoenix 25 8.82
a. Find t h e m is s in g v a lu e s in t h e A N O V A t a b le .
O rla n d o 25 8.83
b. A t t h e 5 % s ig n ific a n c e le v e l, c a n y o u c o n c lu d e
t h a t a v e r a g e c o n t r ib u t io n s d iffe r b y p o litic a l
ANOVA
p a rty ?
Source o f
c. A t t h e 5 % s ig n ific a n c e level, can y o u c o n c lu d e t h a t
Variation SS df MS F p-value F crit
a v e r a g e c o n t r ib u t io n s d iffe r b y in c o m e level?
B e tw e e n G rou ps 4 3 0 .8 7 2 2 1 5 . 4 4 2 0 2 . 9 0 2 .5 8 E - 3 0 3 .1 2 4
4 5 . A t a g y m n a s tic s m e e t , t h r e e ju d g e s e v a lu a t e t h e b a la n c e
W i t h in G r o u p s 7 6 .4 4 72 1 .0 6
b e a m p e r f o r m a n c e s o f fiv e g y m n a s ts . T h e ju d g e s use a scale
o f 1 t o 1 0, w h e r e 1 0 is a p e r f e c t score.
Total 5 07.31 74
Gymnast Judge Means
a. A t t h e 5 % s ig n ific a n c e le v e l, d o m e a n h o u r ly rate s fo r 1 2 3
h o te l m a id s d if f e r b e t w e e n t h e t h r e e cities?
1 8.0 8.5 8.2
b. If d iffe r e n c e s exis t, use T u key's m e t h o d t o d e t e r m in e
2 9.5 9.2 9.7
w h ic h citie s ' m e a n h o u r ly rate s d if f e r a t t h e 5 %
3 7.3 7.5 7.7
s ig n ific a n c e le v e l.
4 8.3 8.7 8.5
4 3 . A n a c c o u n t in g p ro fe s s o r w a n t s t o k n o w if s t u d e n t s p e r f o r m
5 8.8 9.2 9.0
t h e s a m e o n t h e d e p a r t m e n t a l fin a l e x a m irre s p e c tiv e o f
Means
t h e a c c o u n t in g s e c tio n t h e y a t t e n d . T h e f o l lo w in g Excel
A s ta tis tic ia n w a n t s t o e x a m in e t h e o b je c t iv it y a n d a. U se Excel t o g e n e r a t e t h e a p p r o p r i a t e A N O V A ta b le .
c o n s is te n c y o f t h e ju d g e s . S h e p e r f o r m s a t w o - w a y A N O V A b. A t t h e 5 % s ig n ific a n c e le v e l, is t h e r e in t e r a c t io n b e t w e e n
ana ly s is f o r α = 0 .0 1 a n d o b t a in s t h e f o l lo w in g results: f u e l t y p e a n d h y b r id t y p e ?
ANOVA c. A t t h e 5 % s ig n ific a n c e le v e l, c a n y o u c o n c lu d e t h a t
Source o f a v e r a g e f u e l c o n s u m p t io n d iffe rs b y f u e l ty p e ?
Variation SS df MS F p-value F crit
d. A t t h e 5 % s ig n ific a n c e le v e l, c an y o u c o n c lu d e t h a t
Rows 6 .7 4 2 4 1.6 8 6 4 4 .7 5 1.62E -05 7 .0 0 6
a v e r a g e f u e l c o n s u m p t i o n d iffe rs b y t y p e o f h y b rid ?
C olum ns 0 .1 9 2 2 0 .0 9 6 2 .5 5 0 .1 3 9 2 8 .6 4 9
4 7 . A m a n a g e m e n t c o n s u lta n t w a n ts to d e te rm in e w h e th e r
Error 0.301 8 0 .0 3 8
t h e a g e a n d g e n d e r o f a r e s t a u r a n t 's w a i t s t a f f in f lu e n c e
t h e size o f t h e t i p t h e c u s t o m e r le a v e s . T h r e e a g e b r a c k e ts
Total 7 .2 3 5 14
(f a c t o r A in c o lu m n s : y o u n g , m i d d l e - a g e , o ld e r ) a n d g e n d e r
a. A t t h e 1 % s ig n ific a n c e level, can y o u c o n c lu d e t h a t a v e ra g e
(f a c t o r B in ro w s : m a le , f e m a l e ) a r e u s e d t o c o n s t r u c t a
scores d iffe r b y ju d g e ?
r a n d o m b lo c k e x p e r i m e n t . F o r e a c h c o m b i n a t i o n , t h e
b. A t t h e 1 % s ig n ific a n c e le v e l, c a n y o u c o n c lu d e t h a t p e r c e n t a g e o f t h e t o t a l b ill le f t as a t i p f o r 1 0 w a i t s t a f f is
a v e r a g e scores d if f e r b y g y m n a s t ? e x a m i n e d . T h e f o l lo w in g A N O V A t a b l e w i t h α = 0 .0 1 is
c. If a v e r a g e scores d if f e r b y g y m n a s t , use Tu ke y 's H S D pro d u ced .
te s t a t t h e 1 % s ig n ific a n c e le v e l t o d e t e r m in e w h ic h
g y m n a s t s ' p e r f o r m a n c e s d iffer.
ANOVA
36 36 36 If yes, c o n d u c t th e s e te s ts a t t h e 1 % s ig n ific a n c e le v e l.
Hybrid III 43 43 43 If no , e x p la in .
48 48 48
CASE STUDIES
LO 1 4.4 Estimate the m u ltip le linear regression model and in te rp re t the coefficients.
As researchers or analysts, we often need to examine the relationship between tw o or more variables.
We begin this chapter w ith a review o f the correlation coefficient, first discussed in Chapter 3, and
then conduct a hypothesis test to determ ine if tw o variables are significantly correlated. Although
the correlation analysis may establish a linear relationship between tw o variables, it does not d e m
onstrate tha t one variable causes change in the other variable. In this chapter we introduce a m ethod
called regression analysis that assumes tha t one or more variables influence another variable. We
first explore the procedures for estimating a simple linear relationship between tw o variables, co m
m only referred to as the simple linear regression model. We then extend the simple linear regression
model to the case involving several variables, called the m ultiple regression model. Finally, we dis
cuss objective measures to assess how well the estimated model fits the data. These goodness-of-fit
measures are used to select the best-fitting regression model.
I N T R O D U C T O R Y CASE
TABLE 14.1 Income, the U nem ploym ent Rate, and Consumer D ebt Payments, 2 0 1 0 –2 0 1 1
2010; Experian.com collected average m o n th ly consum er d e b t paym ents in A ugust 2010and p ublished the data in N ovem ber 2010.
M a d e ly n w o u ld like t o use t h e s a m p le in fo rm a tio n in T a b le 14.1 to:
2. Use regression analysis to m a k e predictions for d e b t paym ents for given values o f in c o m e and th e u n e m p lo y m e n t rate.
A SCATTERPLOT
A scatterplot is a graphical tool that helps in determining whether or not two vari
ables are related in some systematic way. Each point in the diagram represents a pair
of observed values of the two variables.
Using the data from the introductory case, Figure 14.1 shows a scatterplot depicting the
relationship between income and debt payments. We may infer that the two variables have
a positive relationship; as one increases, the other one tends to increase.
A numerical measure that reveals the direction of the linear relationship between two
variables x and y is called the covariance. It assesses whether a positive or a negative
linear relationship exists between x and y.
where are the sample means of x and y, respectively, and n represents the
number of observations.
A positive value of the sample covariance implies that, on average, when x is above
its mean, y is also above its mean. Similarly, a negative value suggests when x is above its
mean, y is below its mean, indicating a negative linear relationship exists. If the covariance
is zero, then the two variables have no linear relationship. Further interpretation of the
covariance is difficult because it is sensitive to the units of measurement. For instance,
the covariance between two variables might be 100 and the covariance between two other
variables might be 1,000, yet all we can conclude is that both sets of variables are posi
tively related. In other words, we cannot comment on the strength of the relationships.
An easier measure to interpret is the correlation coefficient, which describes both the
direction and strength of the relationship between x and y.
where sx and sy are the sample standard deviations of x and y, respectively, and
− 1 ≤ rxy ≤ 1.
In short, the sample correlation coefficient rxy is unit-free and its value falls between − 1
and 1. If rxy equals 1, then a perfect positive linear relationship exists between x and y.
Similarly, a perfect negative linear relationship exists if rxy equals − 1. If rxy equals zero,
then no linear relationship exists between x and y. Other values for rxy must be interpreted
with reference to − 1 , 0, and 1. As the absolute value of rxy approaches 1, the stronger the
linear relationship. For instance, rxy = − 0.80 indicates a strong negative relationship,
whereas yxr = 0.12 indicates a weak positive relationship. However, we should comment
on the direction of the relationship only if the correlation coefficient is found to be statisti
cally significant— a topic which we address shortly.
EXAMPLE 14.1
Calculate the sample covariance and the sample correlation coefficient between
debt payments and income from the data in Table 14.1. Interpret these values.
Using the overall sum from the third column in Table 14.2 and n = 26, we calculate
the covariance as
Thus, the covariance of 1,119.18 indicates that income and debt payments have a
positive linear relationship. In addition, the correlation coefficient of 0.87 indicates
that the strength of the positive linear relationship is strong. We will soon see that
the correlation coefficient is statistically significant.
Using Excel to Calculate the Covariance
and the Correlation Coefficient
As discussed in Chapter 3, Excel easily produces the covariance and the correlation coef
ficients. Consider the next example.
SO LU TIO N :
A. Open the data labeled Debt Payments found on the text website.
B. Choose Formulas > Insert Function > COVARIANCE.S from the
menu.
C. In the COVARIANCE.S dialog box shown in Figure 14.2, click on the box to
the right of Array 1 and select the income data. Then click on the box to the
right of Array 2 and select the debt payments data. Excel returns a value of
1,119.18.
D. In order to calculate the correlation coefficient, choose Formulas > Insert
Function > CORREL. Select the data as you did when calculating the
covariance. Excel returns a value of 0.87.
Note that we can easily modify the test to a one-tailed test. As in all hypothesis tests, the
next step is to specify and calculate the value of the test statistic.
T E S T S T A T I S T I C F O R p xy
The test statistic for the hypothesis test concerning the significance of the
population correlation coefficient pxy is assumed to follow the t distribution with
d f = n − 2 and its value is calculated as where is
the standard error of rxy. Or, equivalently,
EXAMPLE 14.3
Using the critical value approach to hypothesis testing, determine whether the cor
relation coefficient between income and debt payments is significant at the 5% level.
Using rxy = 0.87 and n = 26 from Example 14.1 and d f = n − 2 = 24, we calculate
the value of the test statistic as
With α = 0.05, tα/2,df = t0.025,24 = 2.064. Thus, the decision rule is to reject H0 if
t24 > 2.064 or t24 < − 2 .064. Since 8.64 > 2.064, we reject H0. At the 5% signifi
cance level, the correlation coefficient between income and debt payments is sig
nificantly different from zero.
If we conduct the test with the p-value approach, we can use the t table to
approximate 2P(T24 ≥ 8.64). Alternatively, we can obtain the exact p -value as 0.0000
using the T.DIST.2T function in Excel (we input X = 8.64 and Deg_freedom = 24).
Consistent with the critical value approach, we reject H0 because the p -value < α .
EXERCISES 14. 1
Concepts a. C a lc u la te t h e v a lu e o f t h e t e s t statistic.
b. A p p r o x im a t e t h e p -v a lu e .
1. Consider th e fo llo w in g sam ple data:
c. A t t h e 5 % s ig n ific a n c e le v e l, w h a t is t h e c o n c lu s io n t o
8 5 3 10 2
x t h e te st? E x p la in .
y 380 210 90 20 2
5. A s a m p le o f 1 0 o b s e r v a t io n s p r o v id e s t h e f o l lo w in g statistics:
a. C o n s tr u c t a n d in t e r p r e t a s c a tte rp lo t.
sx = 1 3 , sy = 18, and sx y = 1 1 7 .2 2
b. C a lc u la te a n d in t e r p r e t t h e s a m p le c o v a ria n c e .
a. C a lc u la te a n d in t e r p r e t t h e s a m p le c o r re la tio n
c. C a lc u la te a n d in t e r p r e t t h e s a m p le c o r re la tio n
c o e ff ic ie n t rxy.
c o e ffic ie n t.
b. S p e c ify t h e h y p o th e s e s t o d e t e r m in e w h e t h e r t h e
2. C o n s id e r t h e f o l lo w in g s a m p le d a ta : p o p u l a t i o n c o r re la tio n c o e f f ic ie n t is po s itiv e .
x -3 0 10 0 23 16 c. C a lc u la te t h e v a lu e o f t h e t e s t s tatistic. A p p r o x im a t e t h e
44 -1 5 -1 0 -2 5 p -v a lu e .
y
a. C o n s t r u c t a n d in t e r p r e t a s c a tte rp lo t. d. A t t h e 5 % s ig n ific a n c e le v e l, w h a t is t h e c o n c lu s io n t o
t h e te st? E x p la in .
b. C a lc u la te a n d in t e r p r e t sxy.
6. A s a m p le o f 2 5 o b s e r v a t io n s p r o v id e s t h e f o l lo w in g statistics:
c. C a lc u la te a n d in t e r p r e t rxy.
sx = 2 , sy = 5 , and sx y = − 1 .7 5
3. C o n s id e r t h e f o l lo w in g c o m p e t in g h y p o th e s e s :
a. C a lc u la te a n d in t e r p r e t t h e s a m p le c o r re la tio n
c o e ff ic ie n t r xy.
b. S p e c ify t h e c o m p e t in g h y p o th e s e s in o r d e r t o d e t e r m in e
Th e sam ple consists o f 25 observations and th e sam ple w h e t h e r t h e p o p u l a t i o n c o r r e la tio n c o e f f ic ie n t d iffers
correlation coefficient is 0.15. f r o m ze ro .
a. Calculate th e value o f th e test statistic. c. M a k e a c o n c lu s io n a t t h e 5 % s ig n ific a n c e le v e l.
b. At th e 5 % significance level, specify th e critical value(s)
an d th e decision rule.
A p p lic a tio n s
c. W h a t is th e conclusion to th e test? Explain.
7. In J u n e 2 0 0 9 a n o n s la u g h t o f m is e r a b le w e a t h e r in N e w
4. Consider th e fo llo w in g c o m p e tin g hypotheses:
E n g la n d p la y e d h a v o c w i t h p e o p le 's p la n s a n d psyches.
H o w e v e r , t h e d r e a r y w e a t h e r b r o u g h t a q u ie t b e n e f it t o
m a n y c ity n e ig h b o r h o o d s . P olice r e p o r t e d t h a t t h e w e a t h e r
T h e s a m p le consists o f 3 0 o b s e r v a t io n s a n d t h e s a m p le w a s a k e y f a c t o r in r e d u c in g f a t a l a n d n o n d e a d ly s h o o tin g s
w h e t h e r t h e p o p u l a t i o n c o r re la tio n b e t w e e n t h e size o f a
June 2009 22 15
h o u s e a n d p r o p e r t y ta x e s d iffe rs f r o m zero .
Source: The Boston Globe, July 3 , 2009.
d. C a lc u la te t h e v a lu e o f t h e t e s t s ta tistic a n d a p p r o x im a t e
a. C a lc u la te a n d in t e r p r e t t h e c o v a r ia n c e a n d t h e its p -v a lu e .
c o r re la tio n c o e ffic ie n t.
e. A t t h e 5 % s ig n ific a n c e le v e l, w h a t is t h e c o n c lu s io n to
b. S p e c ify t h e c o m p e t in g h y p o th e s e s in o r d e r t o d e t e r m in e t h e test?
w h e t h e r t h e r e is a n e g a t iv e p o p u l a t i o n c o r re la tio n
10. F I L E M a n y a t t e m p t s h a v e b e e n m a d e t o r e la te h a p p in e s s
b e t w e e n t h e n u m b e r o f ra in y d a y s a n d c rim e .
w i t h v a rio u s facto rs. O n e such s tu d y relates h a p p in e s s w i t h
c. C a lc u la te t h e v a lu e o f t h e t e s t s ta tis tic a n d a p p r o x im a t e
a g e a n d fin d s t h a t h o ld in g e v e r y t h in g else c o n s ta n t, p e o p le
i t s p -v a lu e .
a re le ast h a p p y w h e n t h e y a re in t h e ir m id - 4 0 s ( The E c o n o m is t,
d. A t t h e 5 % s ig n ific a n c e le v e l, w h a t is t h e c o n c lu s io n t o D e c e m b e r 1 6 , 2 0 1 0 ) . T h e a c c o m p a n y in g t a b le s h o w s a
t h e test? D o e s it a p p e a r t h a t d r e a r y w e a t h e r a n d c rim e p o r t io n o f d a ta o n a re s p o n d e n t's a g e a n d h is /h e r p e r c e p tio n
a re n e g a t iv e ly c o rre la te d ? o f w e ll - b e in g o n a scale f r o m 0 t o 100 ; t h e c o m p le t e d a ta are
8. F I L E D iv e r s if i c a t io n is c o n s id e r e d i m p o r t a n t in f in a n c e o n t h e t e x t w e b s ite , la b e le d H a p p i n e s s a n d A g e .
d a t a o n c lo s in g s to c k p ric e s o f f o u r c o m p a n i e s in 2 0 1 0 .
69 72
T h e e n t i r e d a t a c a n b e f o u n d o n t h e t e x t w e b s i t e , la b e le d
Coca b e t w e e n a g e a n d h a p p in e s s .
Bank of General
M onth Microsoft Cola America Electric b. Is t h e p o p u l a t i o n c o r re la tio n c o e ffic ie n t s tatistica lly
s ig n ific a n t a t t h e 1 % level?
Jan 27.61 49.52 15.13 15.64
c. C o n s tr u c t a s c a t t e r p lo t t o p o in t o u t a f l a w w i t h t h e
Feb 28.22 54.88 16.61 15.72
a b o v e c o r re la tio n analysis.
11. F I L E Th
efo llo w in g t a b le lists t h e N a tio n a l B asketb all
Dec 27.91 55.58 13.34 18.29
A ssociation's le a d in g scorers, t h e ir a v e r a g e p o in ts p e r g a m e
Source: fin a n c e .y a h o o .c o m .
(PPG), a n d t h e ir a v e r a g e m in u te s p e r g a m e (M P G ) f o r 2 0 0 8 ;
a. C o m p u t e t h e c o r re la tio n c o e ffic ie n ts b e t w e e n all pairs o f t h e d a ta a re also a v a ila b le o n t h e t e x t w e b s ite , la b e le d P o in ts .
s to c k prices.
PPG MPG
b. S u p p o s e a n in v e s to r a lr e a d y has a s ta k e in M ic r o s o f t a n d
D. Wade 30.2 38.6
w o u ld like t o a d d a n o t h e r a ss e t t o h e r p o rtfo lio . W h ic h
L. James 28.4 37.7
o f t h e r e m a in in g t h r e e assets w ill g iv e h e r t h e m a x im u m
K. Bryant 26.8 36.1
b e n e f it o f d iv e rs ific a tio n ? (H i n t : F ind t h e a ss e t w i t h t h e
D. Nowitzki 25.9 37.3
lo w e s t c o r re la tio n w i t h M ic r o s o ft.)
D. Granger 25.8 36.2
c. S u p p o s e a n in v e s to r d o e s n o t o w n a n y o f t h e a b o v e f o u r
K. Durant 25.3 39.0
stocks. Pick t w o stocks so t h a t s h e g e ts t h e m a x im u m
(in s q u a re f e e t ) a n d t h e p r o p e r t y ta x e s o w e d b y t h e o w n e r.
C. Bosh 22.7 38.0
b e t w e e n PPG a n d M P G . c o r r e s p o n d in g p -v a lu e .
b. S p e c ify t h e c o m p e t in g h y p o th e s e s in o r d e r t o d e t e r m in e d. A t t h e 5 % s ig n ific a n c e le v e l, w h a t is t h e c o n c lu s io n to
M P G is p o s itiv e .
D E T E R M IN IS T IC VERSUS S T O C H A S T IC R E L A T IO N S H IP S
The relationship between the response variable and the explanatory variables is
deterministic if the value of the response variable is uniquely determined by the
explanatory variables; otherwise, the relationship is stochastic.
Our objective is to develop a mathematical model that accurately captures the relationship
between the response variable y and the k explanatory variables x 1 x 2 , ..., xk. The model
must also account for the randomness that is a part of real life. We start with a determin
istic component that approximates the relationship we want to model, and then make it
stochastic by adding a random error term to it.
In this section we focus on the simple linear regression model, which uses one ex
planatory variable, denoted x b to explain the variability in the response variable, denoted
y. For ease of exposition when discussing the simple linear regression model, we often
drop the subscript on the explanatory variable and refer to it solely as x. In the next section
we extend the simple linear regression model to the multiple regression model, where
more than one explanatory variable is linearly related with the response variable.
A fundamental assumption underlying the simple linear regression model is that the
expected value of y lies on a straight line, denoted by β 0 + β1x, where β0 and β1 (the
Greek letters read as betas) are the unknown intercept and slope parameters, respectively.
The expression β 0 + β1 x is the deterministic component of the regression model, which
can be thought of as the expected value of y for a given value of x. In other words, con
ditional on x, E(y) = β0 + β1 x . The slope parameter β1, determines whether the linear
relationship between x and E(y) is positive (β1, > 0) or negative (β1 < 0); β1 = 0 indicates
that there is no linear relationship. Figure 14.3 shows the deterministic portion of the
regression model for various values of the intercept β0 and the slope β1 parameters.
As noted earlier, the actual value y may differ from the expected value E(y). Therefore,
we add a random error term ε (the Greek letter read as epsilon) to develop a simple linear
regression model.
where y and x are the response variable and the explanatory variable, respectively,
and ε is the random error term. The coefficients β0 and β1 are the unknown param
eters to be estimated.
The sample regression equation for the simple linear regression model is denoted as
We use this equation to make predictions fo r y for a given x . Since the predictions cannot
be totally accurate, the difference between the predicted and the actual value represents
the residual e; that is,
Before estim ating a simple linear regression model, it is extremely useful to
visualize the relationship between y and x by constructing a scatterplot. Here, we
explicitly place y on the vertical axis and x on the horizontal axis, implying that x
influences the variability in y. In Figure 14.1, we used the data from the introductory
case to show a scatterplot of debt payments plotted against income. In Figure 14.4,
we superimpose a linear trend line through the points on the scatterplot. In this chap
ter we focus only on linear relationships; nonlinear relationships will be considered
in Chapter 16.
The superimposed line in Figure 14.4 is the sample regression equation, ŷ = b0 + b 1x,
where y and x represent debt payments and income, respectively. The upward slope of the
line suggests that as income increases, the predicted debt payments also increase. Also,
the vertical distance between any data point on the scatterplot and the corresponding point
on the line, y a n d ŷ, represents the residual, e = y − ŷ .
In essence, b 1 is a ratio of the sample covariance sxy between x and y to the vari
ance s2x of x. This becomes apparent when we rewrite the regression coefficient as
Since we calculate the sample correlation coefficient as
we can express b 1 and rxy in terms of each other.
R E L A T I O N S H I P B E T W E E N b1 A N D r xy
Let b 1 be the estimated slope coefficient of a simple linear regression model
y = β0 + β1 x + ε , and let rxy be the sample correlation coefficient between x and
y. Then
Since sx and sy are non-negative, the signs of b 1 and rxy are always the same.
EXAMPLE 14.4
Using the data from Table 14.1, let debt payments represent the response variable
and income represent the explanatory variable.
a. Calculate and interpret b 1.
b. Calculate and interpret b0.
c. What is the sample regression equation?
d. Predict debt payments if income is $80,000.
Using the summations from the last rows of the last two columns, we com
pute Since we had earlier computed
the sample correlation coefficient as rxy = 0.87 (Example 14.1), we can also
estimate the slope coefficient as (slight
difference due to rounding). As anticipated, the slope is positive, suggesting
a positive relationship between income and debt payments. Since income is
measured in $ 1,000s, our interpretation is that if median household income
increases by $1,000, then on average, we predict consumer debt payments to
increase by b 1, that is, by $10.44.
b . Using b 1 = 10.44 and the sample means of
we obtain an estimate for b0 as b0
= 210.38. This estimated intercept coefficient of 210.38 suggests that if
income equals zero, then predicted debt payments are $210.38. In this par
ticular application, this conclusion makes some sense, since a household
with no income still needs to make debt payments for any credit card use,
automobile loans, and/or a mortgage. However, we should be careful about
predicting y when we use a value for x that is not included in the sample.
In the Debt Payments data set, the lowest and highest values for income (in
$ 1,000s) are $59.40 and $103.50, respectively; plus the scatterplot suggests
that a line fits the data well within this range of the explanatory variable.
Unless we assume that income and debt payments will maintain the same
linear relationship at income values less than $59.40 and more than $103.50
we should refrain from choosing values o f the explanatory variable outside
the sample range.
c. With b0 = 210.38 and b 1 = 10.44, we write the sample regression equation as
SO LU TIO N:
A. Open the data labeled D ebt Payments found on the text website.
B. Choose Data > Data Analysis > Regression from the menu.
C. See Figure 14.5. In the Regression dialog box, click on the box next to Input
Y Range, then select the debt data, including its heading. For Input X Range,
select the income data, including its heading. Check Labels, since we are using
Debt and Income as headings.
D. Click OK.
Regression Statistics
Multiple R 0.8675
R Square 0.7526
Observations 26
ANOVA
df SS MS F Significance F
Total 25 388182.46
EXERCISES 14.2
Concepts a. In t e r p r e t t h e s lo p e c o e ffic ie n t.
b. P r e d i c t y i f x e q u a l s − 15.
1 2 . In a s im p le lin e a r regression, t h e f o llo w in g in fo r m a tio n is g ive n :
1 8. C o n s id e r t h e f o llo w in g s a m p le d a ta :
x 12
23 11 23 14 21 18 16
a. C a lc u la te b1. 28 43 21 40 33 41 37 32
y
b. C a lc u la te b 0.
a. C o n s tr u c t a s c a t t e r p lo t a n d v e r ify t h a t e s t im a t in g a
c. W h a t is t h e s a m p le re g re s s io n e q u a tio n ? P re d ic t y if x
s im p le lin e a r re g re s s io n is a p p r o p r ia t e in th is p r o b le m .
e q u a ls 4 0 .
b. C a lc u la te b1, a n d b 0. W h a t is t h e s a m p le re g re s s io n
13. In a s im p le lin e a r re g re s s io n , t h e f o l lo w in g in fo r m a tio n is
e q u a t io n ?
g iv e n :
c. F ind t h e p r e d ic t e d v a lu e f o r y i f x e q u a ls 1 0 , 15,
and 20.
1 9 . C o n s id e r t h e f o llo w in g s a m p le d a ta :
a. C a lc u la te b1. x
24 27 21 23 14 14 15
b. C a lc u la te b0. 22
101 139 250 88 87 14 16 20
c. W h a t is t h e s a m p le re g re s s io n e q u a t io n ? P re d ic t y if x y
e q u a ls - 2 0 . a. C o n s tr u c t a s c a t t e r p lo t a n d v e r ify t h a t e s t im a t in g a
b. C a lc u la te b 0 a n d b1. W h a t is t h e s a m p le reg re s s io n
e q u a tio n ?
a. C a lc u la te b 1.
c. F in d t h e p r e d ic te d v a lu e f o r y i f x e q u a ls 2 0 , 1 0 0 ,
b. C a lc u la te b 0.
and 200.
c. W h a t is t h e s a m p le re g re s s io n e q u a tio n ? P re d ic t y if x
e q u a ls 2 5 .
A p p lic a t io n s
1 5 . C o n s id e r t h e f o l lo w in g d a ta :
2 0 . T h e d ir e c t o r o f g r a d u a t e a d m is s io n s a t a la r g e u n iv e r s ity
is a n a l y z i n g t h e r e l a t i o n s h i p b e t w e e n s c o re s o n t h e m a t h
a. C a lc u la te r xy. p o r t i o n o f t h e G r a d u a t e R e c o r d E x a m i n a t i o n (G R E ) a n d
b. C a lc u la te b 0. s u b s e q u e n t p e r f o r m a n c e in g r a d u a t e s c h o o l, as m e a s u r e d
c. W h a t is t h e s a m p le re g re s s io n e q u a tio n ? P re d ic t y if x b y a s t u d e n t 's g r a d e p o i n t a v e r a g e (G P A ). S h e u s e s a s a m p l e
e q u a ls - 5 . o f 8 s tu d e n ts w h o g r a d u a te d w it h in t h e p a s t fiv e y e a rs . T h e
d a t a a r e a s f o ll o w s :
16. In a s im p le lin e a r reg re s s io n , t h e f o llo w in g s a m p le re g re s s io n
700
e q u a t io n is o b ta in e d : GRE 720 650 750 680 730 740 780
3.0
ŷ = 1 5 + 2 .5 x . GPA 3.5 3.2 3.7 3.1 3.9 3.3 3.5
a. P r e d i c t y if x e q u a ls 1 0.
a. C o n s tru c t a s c a tte rp lo t p la c in g "GRE" o n t h e h o riz o n ta l axis.
b. W h a t h a p p e n s t o th is p r e d ic t io n if x d o u b le s in v a lu e ?
b. F in d t h e s a m p le re g re s s io n e q u a t io n f o r t h e m o d e l:
17. In a s im p le lin e a r reg re s s io n , t h e f o l lo w in g s a m p le re g re s s io n GPA = β0+β1 GRE + ε .
e q u a t io n is o b t a in e d :
c. W h a t is a s tu d e n t's p r e d ic t e d GPA if h e /s h e s c o re d 7 1 0
19 85 $23,490 $22,887
a. U se Excel to e s tim a te th e m o d e l: S a la ry = β 0 + β 1 ERA + ε
1986 23,866 23,172
a n d in te r p r e t th e c o e ffic ie n t o f ERA.
⋮ ⋮ ⋮
b. U se th e e s tim a te d m o d e l to p r e d ic t s a la ry fo r e a c h
2006 48,398 58,101 p la y e r, g iv e n his ERA. For e x a m p le , use th e s a m p le
w ith ERA = 2 .5 3 .
W h e n e s tim a tin g th e m o d e l: c. D e riv e th e c o rre s p o n d in g re s id u a ls a n d e x p la in w h y th e
C o n s u m p tio n = β 0 + β 1 D is p o s a b le In c o m e + ε , re s id u a ls m ig h t b e so h ig h .
la b e le d H a p p in e s s a n d A g e o n th e te x t w e b s ite .
C o e ffic ie n ts S ta n d a rd t St a t p -v a lue
a. U se Excel to e s tim a te a s im p le re g re s s io n m o d e l w ith
E rro r
593.925 14.397 h a p p in e s s as th e re s p o n s e v a ria b le a n d a g e as th e
In t e r c e p t 8 5 5 0 .6 7 5 5 .1 2 E -1 2
e x p la n a to r y v a ria b le .
D is p o s a b le b. U se th e e s tim a te s to p r e d ic t h a p p in e s s w h e n a g e e q u a ls
In c o m e 0 .6 8 6 0 .0 1 5 4 4 .3 3 2 1.9E-21 2 5 ,5 0 , a n d 7 5 .
The sample regression equation for the multiple regression model is denoted as
S O L U T IO N :
Regression Statistics
Multiple R 0.8676
R Square 0.7527
Adjusted R 0.7312
Square
Observations 26
ANOVA
df SS MS F Significance F
Total 25 388182.46
Standard
Coefficients Error t Stat p-value Lower 95% Upper 95%
4.02 2.57 0.0199 c. Predict the price of a 2,500 square-foot ho m e with three
Intercept 10.35
bedroom s and tw o bathrooms.
Population 8.47 2.71 3.12 0.0062
32. FILE Education reform is one of the most hotly debated
Income 7.62 6.63 1.15 0.2661
subjects on both state and national policy makers' list of
a. What is the sample regression equation? socioeconom ic topics. Consider a regression model that
b. Interpret the slope coefficients. relates school expenditures and family background to
c. Predict sales if a city has 1.5 million w om e n over the age student performance in Massachusetts using 224 school
o f 60 and their average incom e is $44,000. districts. The response variable is the mean score on the
MCAS (Massachusetts Comprehensive Assessment System) statistician wants to use 2009 data to estimate a multiple
exam given in May 1998 to 10th-graders. Four explanatory regression model that links the quarterback's salary with his
variables are used: (1) STR is the student-to-teacher ratio, pass completion percentage (PCT), total touchdowns scored
(2) TSAL is the average teacher's salary, (3) INC is the median (TD), and his age. A portion o f the data is shown below; the
household income, and (4) SGL is the percentage of single com plete data set can be found on the text website, labeled
family households. A portion o f the data is shown below; the Q uarterback Salaries.
com plete data set can be found on the text website, labeled Salary
MCAS. Name (in $ millions) PCT TD Age
TSAL INC Philip Rivers 25.5566 65.2 28 27
Score STR (%) (in $1,000) (in $1,000) SGL (%)
Jay Cutler 22.0441 60.5 27 26
227.00 19.00 44.01 48.89 4.70
230.67 17.90 40.17 43.91 4.60 Tony Romo 0.6260 63.1 26 29
SOURCE: USA Today database for s a la rie s ; NFL.com for other data.
230.67 19.20 44.79 47.64 5.10
a. Estimate the model defined as Salary = β 0 + β 1 PCT +
SOURCE: Massachusetts Department o f Education and the Census o f Population and
Housing.
β 2TD + β 2Age + ε .
b. Are you surprised by the estimated coefficients?
a. For each explanatory variable, discuss whether it is likely
c. Drew Brees earned 12.9895 million dollars in 2009.
to have a positive or negative causal effect on Score.
According to the model, what is his predicted
b. Estimate the sample regression equation. Are the signs
salary given his PCT, TD, and a g e of 70.6,34, and 30,
on the coefficients as expected?
respectively?
c. What is the predicted score if STR = 18, TSAL = 50,
d. Tom Brady earned 8.0073 million dollars in 2009.
INC = 60, SGL = 5?
According to the model, what is his predicted
d. What is the predicted score if everything else is the sam e salary given his PCT,TD, and a g e o f 65.7,28, and 32,
as above except INC = 80? respectively?
33. FILE American football is the highest paying sport on a per e. Compute and interpret the residual salary for Drew Brees
g a m e basis. The quarterback, considered the most important and Tom Brady.
player on the team, is appropriately com pensated. A sports
A numerical measure that gauges dispersion from the sample regression equation is the
sample variance of the residual, denoted s2e. We generally report the standard deviation of
the residual, denoted se, more commonly referred to as the standard error of the esti
mate. The variance s2eis defined as the average squared difference between a n d ŷi. The
numerator of the formula is the error sum of squares SSE = ∑ (yi − ŷi) 2 = ∑ e 2i. Dividing
SSE by its respective degrees of freedom n − k − 1 yields s2e, also referred to as the mean
square error M S E . Recall that k denotes the number of explanatory variables in the
regression model; thus, for a simple linear regression model, k equals one. The standard
error of the estimate se is the positive square root of s2e. The less the dispersion, the smaller
the se, which implies a better fit to the model. As mentioned earlier, if all points fall on the
line, then no dispersion occurs, so se equals zero.
The standard error of the estimate se is a point estimate of the standard deviation
of the random error ε , and is calculated as
Theoretically, se. can assume any value between zero and infinity, 0 ≤ se < ∞ . The
closer se is to zero, the better the model fits.
EXAMPLE 14.7
Consider the sample data in Table 14.1 and the regression output for Model 1 in
Table 14.4. Use the sample regression equation, Income, to
calculate and interpret the standard error of the estimate se.
S O L U T IO N : First, we calculate the variance of the residual s2e
. Let y and x denote
Debt and Income, respectively. The first two columns of Table 14.6 show the values
of these variables. The third column shows the predicted values ŷ and the fourth
column shows the squared residuals, e2 = (y − ŷ )2. The value in the last row of the
last column is the error sum of the squares SSE and is the numerator in the formula
for se2.
The standard error of the estimate is measured in the same units of measurement as
the response variable. Since debt payments are in dollars, we report se as $63.26.
We mentioned earlier that the closer se is to zero, the better the fit. Can we con
clude that a value of $63.26 is close to zero? A general guideline is to evaluate the
value of se with respect to the mean of the response variable. If we construct the
ratio and find that this ratio is less than 0.20, then we have some evidence that
the model is an effective analytical and forecasting tool. Given that average debt
payments are $983.46, we compute the ratio as 63.26/983.46 = 0.06. Thus, Model
1 appears promising.
Most of the time we rely on statistical software packages to report se. (If se is not explicitly
given, other statistics like SSE are generally provided, which then greatly facilitate the
calculation of se.) Excel reports the value for se in the regression output section entitled
Regression Statistics. It is simply referred to as Standard Error. In column 2 of Table 14.7,
we report the Excel regression statistics for Model 1. Note that se = 63.26 is the same as
the one calculated above.
TABLE 14.7 Regression Statistics for Model 1 and Model 2
Model 1 Model 2
Multiple R 0.8675 0.8676
R Square 0.7526 0.7527
Adjusted R Square 0.7423 0.7312
Standard Error 63.26 64.61
Observations 26 26
Regression Equation ŷ = 210.30 + 10.44x ŷ = 199 + 10.51x1 + 0.62x2
Our objective in adding another explanatory variable to the regression model is to in
crease the model’s usefulness. In Model 2 we use income x, and the unemployment rate
x 2 to explain debt payments y. If Model 2 is an improvement over Model 1, then we
would expect it to have a smaller standard error of the estimate. Table 14.7 also shows the
relevant regression statistics for Model 2. We note that the standard error of the estimate
for Model 2, se = $64.61, is actually greater than that for Model 1 (64.61 > 63.26). In
other words, there is less dispersion between the observed values of debt payments and
the predicted values of debt payments when we include only one explanatory variable in
the model. So far, this suggests that Model 1 provides a better fit for the sample data. In
general, we use the standard error of the estimate in conjunction with other measures to
judge the overall usefulness of a model.
For ease of exposition, we show a scatterplot with all the points removed except one
(point A ). Point A refers to the observation (xi, y i) . The blue line represents the estimated
regression equation based on the entire sample data; the horizontal and vertical green
lines represent the sample means respectively. The vertical distance (difference)
between the data noint A and (point C) is the total variation (distance A C ). For
each data point, we square these differences and then find their sum— this amounts to
Now, we focus on the difference between the predicted value of the data point ŷi
(point B ) and This difference is the explained variation in y (distance BC ). Explained
variation indicates that the positive difference between occurs because y rises with
x and in this case x is above its mean. Squaring all such differences and summing them
yields the regression sum of squares.
The difference between the particular observation and its predicted value (dis
tance AB ) is the unexplained variation in y. This is the portion of the variation in y that
remains unexplained; it is the variation due to random error or chance. Squaring all
such differences and summing them yields the familiar error sum of squares, SSE =
∑(yi − ŷi)2.
Thus, the total variation in y can be decomposed into explained and unexplained varia
tion as follows:
SST = SSR + SSE
The above equation represents the coefficient of determination R2. The value of R2 falls
between zero and one, 0 ≤ R2 ≤ 1. The closer R 2 is to one, the stronger the fit; the closer
it is to zero, the weaker the fit.
T H E C O E F F I C I E N T O F D E T E R M I N A T I O N R2
where
Most statistical packages, including Excel, provide the ANOVA calculations for SST,
SSR, and SSE.
EXAMPLE 1 4 .8
Calculate and interpret the coefficient of determination R2 given the sample data
in Table 14.1 and the sample regression equation from Model 1:
10.44Income.
Therefore,
Note that this value matches the Excel estimate shown in Table 14.7. The coefficient
of determination R 2 shows that 75.26% of the sample variation in debt payments is
explained by changes in income.
Another interesting statistic in Table 14.7 is Multiple R . Using our earlier notation, this
measure is simply the sample correlation between the response variable y and its pre
dicted value ŷ, or ryŷ. In a simple linear regression model, Multiple R also represents the
absolute value of the correlation between y and x , or |r yx|. Moreover, the coefficient of
determination is the square of Multiple R, that is, R2 = ryŷ2.
SOLUTION: The Multiple R in Table 14.7 implies that the sample correlation
coefficient between y and ŷ for Model 2 is ry ŷ = 0.8676. Thus, R2 =ŷr2y = 0.86762 =
0.7527.
Recall that the standard error of the estimate for Model 1 (se = 63.26) was smaller than
that for Model 2 (se = 64.61), suggesting that Model 1 provides a better fit. Now, the
coefficient of determination for Model 2 (R2 = 0.7527) is slightly higher than that of
Model 1 (R2 = 0.7526), implying that Model 2 explains more of the variation in debt
payments. How do we resolve these apparent conflicting results? It turns out that we can
not use R2 for model comparison when the competing models do not include the same
number of explanatory variables. This occurs because R2 never decreases as we add more
explanatory variables to the model. A popular model selection criterion in such situations
is adjusted R 2.
A D J U S T E D R2
If SSE is substantially greater than zero and k is large compared to n, then adjusted R2 will
differ substantially from R2. Adjusted R2 may be negative, if the correlation between the
response variable and the explanatory variables is sufficiently low.
EXAMPLE 1 4 .1 0
Using the regression statistics from Table 14.7, use the value of the adjusted R2 for
model comparison.
A r e c e n t s t u d y s h o w s s u b s t a n t ia l v a r i a b i l i t y in c o n s u m e r
d e b t p a y m e n t s d e p e n d i n g o n w h e r e t h e c o n s u m e r re s id e s
( E x p e r ia n .c o m , N o v e m b e r 1 1 , 2 0 1 0 ) . A p o s s ib le e x p la n a t io n
is t h a t a lin e a r r e l a t io n s h ip e x is ts b e t w e e n c o n s u m e r d e b t
p a y m e n t s a n d a n area's m e d ia n h o u s e h o ld in c o m e . In o r d e r
t o s u b s t a n t ia t e t h is c la im , r e le v a n t d a t a o n 2 6 m e t r o p o l i
t a n a re a s a re c o lle c t e d . T h e c o r r e la t io n c o e f f ic ie n t b e t w e e n
d e b t p a y m e n t s a n d i n c o m e is c o m p u t e d as 0 .8 7 , s u g g e s t
in g a s t r o n g p o s it iv e lin e a r r e l a t io n s h ip b e t w e e n t h e t w o
v a ria b le s . A s im p le t e s t c o n f ir m s t h a t t h e c o r r e la t io n c o e f
f ic ie n t is s t a t is t ic a lly s ig n if ic a n t a t t h e 5 % le v e l.
T w o r e g r e s s io n m o d e l s a r e a ls o e s t i m a t e d f o r t h e a n a ly s is . A s i m p l e lin e a r r e
g r e s s io n m o d e l ( M o d e l 1), u s in g c o n s u m e r d e b t p a y m e n t s as t h e r e s p o n s e v a r i a b le
a n d m e d i a n h o u s e h o l d in c o m e as t h e e x p l a n a t o r y v a r i a b le , is e s t i m a t e d as
2 1 0 .3 0 + 1 0 . 4 4 ln c o m e . F o r e v e r y $ 1 , 0 0 0 in c r e a s e in m e d i a n h o u s e h o l d in c o m e , c o n
s u m e r d e b t p a y m e n t s a r e p r e d i c t e d t o in c r e a s e b y $ 1 0 .4 4 . In a n a t t e m p t t o im p r o v e
u p o n t h e p r e d i c t i o n , a m u l t i p l e r e g r e s s io n m o d e l ( M o d e l 2 ) is p r o p o s e d , w h e r e m e d ia n
h o u s e h o l d in c o m e a n d t h e u n e m p l o y m e n t r a t e a r e u s e d as e x p l a n a t o r y v a r ia b le s . T h e
s a m p le r e g r e s s io n lin e f o r M o d e l 2 is
G i v e n its s l o p e c o e f f i c i e n t o f o n l y 0 . 6 2 , t h e e c o n o m i c i m p a c t o f t h e u n e m p l o y m e n t
rate o n c o n s u m e r d e b t p a y m e n t s , w ith m e d i a n h o u s e h o l d i n c o m e h e ld fixed , s e e m s
e x t r e m e l y w e a k . G o o d n e s s - o f - f i t m e a s u r e s c o n f i r m t h a t M o d e l 1 p r o v i d e s a b e t t e r fit
t h a n M o d e l 2. T h e s t a n d a r d e rror o f t h e e s t i m a t e is s m a l l e r f o r M o d e l 1, s u g g e s t i n g
l e s s d i s p e r s i o n o f t h e d a t a f r o m t h e s a m p l e r e g r e s s i o n e q u a t i o n . In a d d i t i o n , t h e a d
j u s t e d R2 is h i g h e r fo r M o d e l 1, i m p l y i n g t h a t m o r e o f t h e v a r ia b il i t y in c o n s u m e r d e b t
p a y m e n t s is e x p l a i n e d b y t h e s i m p l e r e g r e s s i o n m o d e l . U s i n g M o d e l 1 a n d a n area's
m e d ia n h o u s e h o l d i n c o m e o f $ 8 0 ,0 0 0 , c o n s u m e r d e b t p a y m e n t s are p r e d ic te d t o b e
$ 1 ,0 4 5 .5 0 .
EXERCISES 14.4
a. Calculate the standard error of the estimate. Given that ANOVA df SS MS F Significance F
the mean of the response variable is 345, d o es the model Regression 2 918.746 459.373 2.817 0.0774
seem promising? Explain. 27 4402.786 163.066
Residual
b. Calculate and interpret the coefficient of determination.
Total 29 5321.532
c. Calculate adjusted R2.
a. C a lc u la te th e s ta n d a rd e rro r o f th e e s tim a te . G iv e n th a t
Applications th e m e a n re tu rn fo r th is s a m p le is 1.31% , c a lc u la te a n d
trying to determine whether a student's high school GPA or b. C alculate and in te rp re t th e co e ffic ie n t o f d e te rm in a tio n .
SAT score is a better predictor o f the student's subsequent c. Calculate th e co rre sp o n d in g a d justed R2.
college GPA. She formulates tw o models:
43. FILE Is it defense or offense that wins football games?
Model 1. College GPA = β0 + β1 High School GPA + ε Consider the following data, which include a team's winning
Model 2. College GPA = β0 + β1 ,SAT Score + ε record (Win), the average number of yards made, and the
average number of yards allowed during the 2009 NFL
She estim ates th e se models using data from a sample of
season. The com plete data, labeled Football, can be found on
10 recent college graduates. A portion of the results are as
the text website.
follows:
ANOVA Results for M odel 1 Team Win (%) Yards M ade Yards Allowed
Regression 1 1.4415 1.4415 11.5032 0.0095 Atlanta Falcons 56.30 340.40 348.90
ANOVA df SS MS F Significance F of the 455 highest-paid CEOs in 2006 with two performance
measures (industry-adjusted return on assets (Adj ROA) and
Regression 2 3549788 1774894 16.12513 8.5E-06
industry-adjusted stock return (Adj Stock Return)) and the
Residual 38 4182663 110070.1
firm's size (Total Assets). The complete data, labeled Executive
Total 40 7732451 Compensation, are available on the text website.
Compensation Adj Stock Total Assets a. Estimate three simple linear regression models that use
(in $ millions) Adj ROA Return (in $ millions) compensation as the response variable with Adj ROA,
TABLE 1 4 . 8 Monthly Return Data for Five Mutual Funds, January 2001 - D e c e m b e r 2010 FILE
Matthew would like to use the sample information in Table 14.8 to:
1. Calculate and interpret the sample correlation coefficient of each fund with
Vanguard’s 500 Index.
2. Make a recommendation for a mutual fund that is not correlated with Vanguard’s
500 Index.
In attempting to create a well-diversified portfolio, an analysis of the correlation between
Sample assets’ returns is crucial. The correlation coefficient measures the direction and the
strength of the linear relationship between assets’ returns. This statistic can aid in the
Report- hunt for assets that form a portfolio. An investor has already chosen Vanguard's 500 Index
mutual fund as part of her portfolio. When choosing to add to her portfolio, she considers
Making these four mutual funds from the Vanguard family:
Investment • Inflation-Protected Securities Index
• Intermediate-Term Bond Index
Decisions by • Real Estate Investment Trust Index
Diversifying • Small Cap Index
Ten years of monthly return data for each of these prospective funds, as well as the
500 Index, are collected. The first row of Table 14.A shows the sample correlation coef
ficients between the 500 Index and each mutual fund.
Correlation
Coefficient 0.0796 − 0.0408 0.6630 0.9030
Test Statistic 0.87 − 0.44 9.62 22.83
p-value 0.39 0.66 0.00 0.00
The correlation coefficient always assumes a value between − 1 and 1; an absolute value
close to 1 implies that the two assets move in sync. In this sample, the highest sample
correlation coefficient is between the 500 Index and the Small Cap Index, with a value of
0.9030. Next on the list is the correlation of 0.6630 between the 500 Index and the Real
Estate Index. Sometimes investors choose assets that are negatively correlated in order
to hedge risk; that is, one asset does well when the other does poorly. Such strategies are
only recommended in the short term, since returns essentially cancel one another out. The
investor wants to invest across a range of asset classes that earn respectable returns but
are relatively uncorrelated. This way, if one asset in a portfolio suffers, the rest may be
unaffected. Given that the Inflation-Protected Securities Index and the Intermediate-Term
Bond Index have correlation coefficients close to zero, these may prove to be desirable
additions to the investor’s portfolio.
A hypothesis test is conducted in order to determine whether the population correla
tion coefficients are significantly different from zero at the 5% significance level. The
null hypothesis is that the returns are uncorrelated and the alternative hypothesis suggests
either a positive or a negative correlation. Rows 2 and 3 of Table 14. A show the value of
the test statistics and the corresponding p -values. For instance, given a p -value of 0.39, the
correlation coefficient between the 500 Index and the Inflation-Protected Securities Index
is not significantly different from zero at the 5% significance level. This same conclusion
holds for the correlation coefficient between the 500 Index and the Intermediate-Term
Bond Index. On the other hand, given p -values of 0.00 for both of the test statistics as
sociated with the correlation coefficient between the 500 Index and the Real Estate Index
and the 500 Index and the Small Cap Index, we can conclude that these correlation coef
ficients are significantly different from zero.
Assuming that the correlation between assets is likely to remain stable in the future,
then it appears that the investor should add either the Inflation-Protected Securities Index
or the Intermediate Bond Index to her portfolio. Compared to the other two funds, these
funds would offer the maximum benefit from diversification in the sense of reducing
volatility.
Conceptual Review
LO 14.1 Conduct a hypothesis test for the population correlation coefficient.
The covariance is a measure of the linear relationship between two variables x and
y. We compute the sample covariance as where n represents the
number of observations. The correlation coefficient is a unit-free measure that gauges
the strength of the linear relationship between two variables x and y. We calculate the
sample correlation coefficient as
When determining whether the correlation coefficient differs from zero, the null
and alternative hypotheses are formulated as H0: ρ xy = 0 and HA: ρxy + 0, where ρxy
is the population correlation coefficient; one-tailed tests are constructed similarly.
The corresponding test statistic for conducting the test follows the t distribution with
d f = n − 2, and its value is calculated as
LO 1 4.3 Estimate the simple linear regression model and interpret the coefficients.
Regression analysis explicitly assumes that one variable, called the response variable,
is influenced by other variables, called the explanatory variables.
The simple linear regression model uses only one explanatory variable to pre
dict and/or describe change in the response variable. The model is expressed as
y = β 0 + β 1 x + ε , where y and x are the response variable and the explanatory vari
able, respectively, and ε is the random error term. The coefficients β0 and β1 are the
unknown parameters to be estimated.
We apply the ordinary least squares (OLS) method to find a sample regression
equation, ŷ = b0 + b 1X, where ŷ is the predicted value of the response variable and b0
and b 1 are the point estimates of β 0 and β1 , respectively. The estimated slope coef
ficient b 1 represents the change in ŷ when x changes by one unit. The units of b 1 are
the same as those of y.
LO 1 4 .4 Estimate the m ultiple linear regression model and interpret the coefficients.
The multiple regression model allows more than one explanatory variable to be lin
early related with the response variable y. It is defined as y = β 0 + β1x1 + β2x2 + . . . +
β kxk + ε , where y is the response variable, x 1 , x 2 , xk are the k explanatory variables
and ε is the random error term. The coefficients β0 , β1 , ..., βk are the unknown param
eters to be estimated. We again use the OLS method to arrive at the following sample
regression equation: ŷ = b0 + b 1x 1 + b2x 2 + . . . + bkx k, where b0, b 1, . . . , bk are the
point estimates of β0 , β 1 , .. ., β k , respectively.
For each explanatory variable xj (j = 1 , . . . , k ), the corresponding slope coefficient bj
is the estimated regression coefficient. It measures the change in the predicted value
of the response variable ŷ given a unit increase in the associated explanatory variable
xj, holding all other explanatory variables constant. In other words, it represents the
partial influence of xj on ŷ .
LO 1 4.5 Calculate and interpret the standard error of the estimate.
The standard error of the estimate se is a point estimate of the standard deviation
of the random error ε , and is calculated as
Theoretically, se can assume any value between zero and
infinity, 0 ≤ se < ∞; the closer se is to zero, the better the model fits. Since se has no
predetermined upper limit, it is difficult to interpret the standard error of the estimate in
isolation; however, if the ratio is less than 0.20, then this indicates that the model
is an effective analytical and forecasting tool.
Year Energy x Healthcare y dem and dynamics, as well as longer-term interest rate
expectations. Consider the following annualized rates for
2004 36.65 9.51
3-month Treasury yields and 10-year Treasury yields.
2005 44.60 15.41
Year 3-M o nth Yield (%) 10-Year Yield (%)
2006 19.68 10.87
2001 3.47 5.02
2007 37.00 4.43
2002 1.63 4.61
2008 - 4 2 .8 7 − 18.45
2003 1.03 4.02
2004 1.40 4.27
sx = 35.77 sy= 13.34
2005 3.21 4.29
sxy= 447.68
2006 4.85 4.79
Source: w w w . fin a n c e . y a h o o .c o m .
2007 4.47 4.63
a. Calculate and interpret the sample correlation coefficient rxy.
2008 1.39 3.67
b. Specify the com peting hypotheses in order to determine
2009 0.15 3.26
whether the population correlation coefficient is
2010 0.14 3.21
significantly different from zero.
SOURCE: Federal Reserve Bank o f Dallas.
c. At the 5% significance level, what is the conclusion
to the test? Are the returns on th e mutual funds a. Construct and interpret a scatterplot o f a 10-year
significantly correlated? treasury yield against a 3-m onth yield.
b. Calculate and interpret the sample correlation a. Estimate: Return = β 0 + β 1 P /E + β 2 P / S + ε . Are t h e
coefficient. Use α = 0.05 to test if the population sig n s on t h e coefficien ts as e x p e c t e d ? Explain.
correlation coefficient is significantly different
b. Interpret t h e s l o p e coeffic ien t o f t h e P / S ratio.
from zero.
c. What is t h e predicted return for a firm with a P /E ratio of
c. Estimate and interpret a sample regression equation
10 an d a P / S ratio o f 2?
using the 10-year yield as the response variable and the
d. What is t h e standard error o f t h e estim a te ? Calculate and
3-month yield as the explanatory variable.
interpret
47. FILE The h om eow n ersh ip rate in the U.S. was 67.4% in
e. In t e r p r e t R2.
2009. In order to determ ine if hom eow n ersh ip is linked
49. FILE There has b een a lot o f discussion regarding the
with income, 2009 state level data on hom eow n ersh ip
relationship b e tw e en Scholastic Aptitude Test (SAT) scores
rate (Ownership) and median household incom e (Income)
and test-takers'family incom e (New York Times, August 27,
were collected. A portion o f th e data is sh ow n below; the
2009). It is generally believed that the wealthier a student's
c om plete data can be found on t h e text website, labeled
family, th e higher th e SAT score. Another com m only
Home Ownership.
used predictor for SAT scores is the student's grade point
State Income Ownership average (GPA). Consider the following data collected on 24
Alabama $39,980 74.1% students. The data can also be found on t h e text website,
labeled SAT.
Alaska $61,604 6 6 .8 %
SAT Income GPA
a. E s tim a te a n d in te rp re t th e m o d e l: O w n e rs h ip = β 0 +
β1 ,In c o m e + ε .
1940 113,000 3.96
1 $154.38 3 9
30. Walt Disney Company 16.3 13.94 1.94
SOURCE: Association o f University Managers and National Science Foundation.
SOURCE: The 2010 returns (January 1, 2010-December 31, 2010) were obtained from
The Wall Street Journal, January 3 , 2010; the P/E ratios and the P/S ratios were obtained a. Estimate: Startups = β 0 + Research + β 2 Patents +
from finance.yahoo.com on January 2 0 , 2011. β 3 Duration + ε .
b. Predict the number of startups for a university that spent (EXPER), and age (AGE). The data can be found on the text
$120 million on research, issued 8 patents, and has had a website, labeled Hourly Wage.
technology transfer office for 20 years. a. Estimate: Wage = β 0 + β 1 EDUC + β 2EXPER + β 3AGE + ε .
c. How much more research expenditure is n eeded for the b. Are the signs as expected?
university to have an additional predicted startup, with c. Interpret the coefficient of EDUC.
everything else being the same?
d. Interpret the coefficient of determination.
51. F I L E researcher interviews 50 em ployees of a large e. Predict the hourly w a g e o f a 40-year-old em plo y ee w h o
manufacturer and collects data on each worker's hourly has 5 years o f higher education and 8 years of experience.
w a g e (Wage), years of higher education (EDUC), experience
CASE STUDIES
SOURCE FORT HE U NEMPLOYMENT RATE D ATA: Development Department, State o f California, June 2009.
LEARNING OBJECTIVES
E R
Inference with
Regression Models
LO 1 5.1 Conduct tests of individual significance.
After reading
this chapter
you should LO 1 5.2 Conduct a test of jo in t significance.
be able to:
LO 1 5.3 Conduct a general test of linear restrictions.
In Chapter 14 we employed simple and m ultiple regression models to find a relationship between
a response variable and one or more explanatory variables. We also studied objective goodness-of-
fit measures that assess how well the sample regression eguation fits the data. While the estimated
regression models and goodness-of-fit measures are useful, it is not clear if the conclusions based
on the estimated coefficients are real or due to chance. In this chapter w e focus on statistical infer
ence w ith regression models. In particular, we develop hypothesis tests tha t enable us to determine
the individual and jo in t significance o f the explanatory variables. We also develop interval estimates
for a prediction from the sample regression equation. Finally, we examine the im portance o f the as
sumptions on the statistical properties o f the OLS estimator, as well as the validity o f the testing pro
cedures. We address com m on violations to the model assumptions, the consequences w hen these
assumptions are violated, and offer some remedial measures.
I N T R O D U C T O R Y C A S E
TABLE 15.1 Winning Percentage, Batting Average, and Earned Run Average in Baseball
T h r e e r e g r e s s i o n m o d e l s are e s t i m a t e d . M o d e l 1 p r e d i c t s w i n n i n g p e r c e n t a g e b a s e d o n BA, w h e r e a s M o d e l 2 u s e s
ERA; M o d e l 3 a p p l i e s b o t h BA a n d ERA t o p r e d i c t w i n n i n g p e r c e n t a g e .
U s e t h e s a m p l e i n f o r m a t i o n to:
1. E m p l o y g o o d n e s s - o f - f i t m e a s u r e s t o d e t e r m i n e w h i c h o f t h e t h r e e m o d e l s b e s t fits t h e data.
2. D e t e r m i n e t h e i n d iv id u a l a n d j o i n t s i g n i f i c a n c e o f BA a n d ERA at t h e 5% s i g n i f i c a n c e level.
A s y n o p s i s o f th is c a s e is p r o v i d e d at t h e e n d o f S e c t i o n 15.1.
15.1 Tests of Significance
This section continues from Chapter 14 with the assessment of linear regression models.
Here we turn our attention to hypothesis testing about the unknown parameters (coef
ficients) of the population regression model. In particular, we test the regression coef
ficients both individually and jointly in order to determine whether there is evidence of a
linear relationship between the response and the explanatory variables. We must point out
that these tests require that the OLS estimators be normally distributed. This condition is
satisfied if the random error term of the model has a normal distribution. If we cannot as
sume that the random error is normal, then the tests are valid only for large sample sizes.
The objective outlined in the introductory case study is to predict a baseball team’s
winning percentage, denoted Win, either on the basis of its batting average BA or its
earned run average ERA, or jointly by BA and ERA. For those readers who do not follow
baseball, BA is a ratio of hits divided by times at bat, and ERA is the average number
of earned runs given up by a pitcher per nine innings pitched. A priori, we expect that a
large BA positively influences a team’s winning percentage, while a large ERA negatively
affects a team’s winning percentage. We define
Model 1 as Win = β 0 + β 1 BA + ε ,
Model 2 as Win = β0 + β 1 ERA + ε , and
Model 3 as Win = β0 + β1 BA + β 2 ERA + ε .
Before we develop the tests of significance, we will use goodness-of-fit measures dis
cussed in Chapter 14 to choose the appropriate model. Table 15.2 shows the relevant
Excel output for the three models; we advise you to replicate these results using the
sample data in Table 15.1.
TABLE 15.2 Relevant Regression Output to Compare th e Models
We choose Model 3 to predict winning percentage because it has the lowest standard error
of the estimate and the highest adjusted R2.
H 0: β j = β j 0 H 0: βj ≤ βj0 H 0: βj ≥ βj0
H 0: βj ≠ βj0 H 0: βj > βj0 H A : βj < βj0
When testing whether xj significantly influences y , we set β j0 = 0 and specify a two-tailed
test as H0: βj = 0 and HA: βj ≠ 0. We could easily specify one-tailed competing hypoth
eses for a positive linear relationship (H0: β j ≤ 0 and HA: β j > 0) or a negative linear
relationship (H0: βj ≥ 0 and HA: βj < 0).
Although most tests of significance are based on β j 0 = 0, in some situations we might
wish to determine whether the slope coefficient differs from a nonzero value. For instance,
we may want to determine if an extra hour of review before the exam will increase a
student’s score by more than 5 points. Here we formulate the hypotheses as H0: βj, ≤ 5
and HA: βj > 5 where βj0 = 5. The procedure for conducting the test remains the same.
As in all hypothesis tests, the next step is to specify the test statistic and compute its
value.
The test statistic for a test of individual significance is assumed to follow the tdf
distribution with d f = n − k − 1 and its value is
where n is the sample size, k is the number of explanatory variables, sbj is the stan
dard error of the OLS estimator bj and βj 0 is the hypothesized value of β j . If βj0=
0,
the value of the test statistic reduces to
EXAMPLE 15.1
Let’s revisit Model 3, Win = β0 + β1 BA + β 2 ERA + ε , estimated with the sample
data in Table 15.1. Conduct a hypothesis test to determine whether batting average
influences winning percentage at the 5% significance level.
SOLUTION: We will use the p -value approach for the test. We set up the following
hypotheses in order to determine whether winning percentage and batting average
have a linear relationship:
H 0: β1 = 0
H a: β 1 ≠ 0
Table 15.3 shows a portion of the regression output that is needed to conduct the
tests of individual significance.
From Table 15.3, we find that b 1 = 3.2754 and sb1 = 0.6723. In addition, given
n = 30 and k = 2, we find d f = n − k − 1 = 30 − 2 − 1 = 27. So we calculate the
value of the test statistic as From Table 15.3
we see that this calculation is not necessary since Excel automatically provides the
value of the test statistic and its associated p -value.
We can use the results reported by Excel since we have a standard case where
β j0 = 0; shortly, we will see an application with a nonstandard case. As usual, the
decision rule is to reject H0 if the p -value < α . Since the p -value is approximately
zero, we reject H0. At the 5% significance level, there is a linear relationship be
tween winning percentage and batting average; in other words, batting average is
significant in explaining winning percentage.
C O M P U T E R - G E N E R A T E D T E ST S T A T IS T IC A N D T H E P- V A L U E
Excel and virtually all other statistical packages report a value of the test statistic
and its associated p -value for a two-tailed test that assesses whether the regression
coefficient differs from zero.
• If we specify a one-tailed test, then we need to divide the computer-generated
p -value in half.
• If we test whether the coefficient differs from a nonzero value, then we cannot
use the value of the computer-generated test statistic and p -value.
C O N F ID E N C E INTERVAL FOR β j
A 100(1 − α )% confidence interval of the regression coefficient β j is computed as
Excel automatically provides a 95% confidence interval for the regression coefficients; it
will provide other levels if prompted. In general, if the confidence interval for the slope
coefficient contains the value zero, then the explanatory variable associated with the
regression coefficient is not significant. Conversely, if the confidence interval does not
contain the value zero, then the explanatory variable associated with the regression coef
ficient is significant.
EXAMPLE 15.2
Use a confidence interval to determine whether earned run average is significant in
explaining winning percentage at the 5% significance level.
SOLUTION: For testing whether earned run average is significant, we set up the
following hypotheses:
H 0: β 2 = 0
H A: β 2 ≠ 0
EXAMPLE 15.3
Johnson & Johnson (J&J) was founded more than 120 years ago on the premise
that doctors and nurses should use sterile products to treat people’s wounds. Since
that time, J&J products have become staples in most people’s homes. Consider the
CAPM where the J&J risk-adjusted stock return R − R f is used as the response vari
able and the risk-adjusted market return R M − Rf is used as the explanatory variable.
A portion of 60 months of data is shown in Table 15.4; the complete data can be
found on the text website, labeled Johnson & Johnson.
I FILE Date R -R f Rm — Rf
a. Since consumer staples comprise many of the products sold by J&J, its stock
is often considered less risky; that is, people need these products whether the
economy is good or bad. At the 5% significance level, is the beta coefficient
less than one?
b. At the 5% significance level, are there abnormal returns? In other words, is the
alpha coefficient significantly different from zero?
SOLUTION: We use the critical value approach for the tests. Using the CAPM
notation, we estimate the model, R − Rf = α + β ( RM− Rf) + ε ; the relevant portion
of the Excel output is presented in Table 15.5.
TABLE 15.5 Portion o f CAPM Regression Results for J&J
a. The estimate for the beta coefficient is 0.5844 and its standard error is
0.0803. Interestingly, our estimate is identical to the beta reported in the
popular press (www.dailyfinance.com, March 4, 2011). In order to determine
whether the beta coefficient is significantly less than one, we formulate the
hypotheses as
H0: β ≥ 1
H a: β < 1
H 0: β1 = β2 = . . . = βk = 0
HA: At least one βj≠ 0.
You might be tempted to implement this test by performing a series of tests o f indi
vidual significance with the t statistic. However, such an option is not appropriate.
The test of joint significance determines if at least one of the explanatory variables
is significant. Therefore, it is not clear if one or all of the explanatory variables must
be significant in order to document a joint significance. In addition, recall from the
ANOVA chapter that if we conduct many individual tests at a given α , the resulting
significance level for the jo int test will not be the same.
Testing a series of individual hypotheses is not equivalent to testing the same
hypotheses jointly.
To conduct the test of joint significance, we employ a one-tailed F test. (Recall that the
distribution introduced in Chapter 11 was used for hypothesis testing in Chap
ters 11 and 13.) The test statistic measures how well the regression equation explains
the variability in the response variable. It is defined as the ratio of the mean regression
sum of squares MSR to the mean error sum of squares MSE, where MSR = SSR /k and
M SE = SSE/(n − k − 1). These values, including the test statistic, are provided in
the ANOVA portion of the regression results.
The test statistic for a test of joint significance is assumed to follow the dis
tribution with d f 1= k and df 2 = n − k − 1 and its value is
where MSR and MSE are the mean regression sum of squares and the mean error
sum of squares, respectively.
EXAMPLE 1 5 .4
Let’s revisit Model 3, Win = β 0 + β 1 BA + β 2ERA + ε , estimated with the sample
data in Table 15.1. Conduct a test to determine if batting average and earned run
average are jointly significant in explaining winning percentage at α = 0.05.
SOLUTION: When testing whether the explanatory variables are jointly significant
in explaining winning percentage, we set up the following hypotheses:
H0: β 1 = β 2 = 0
HA: At least one β j ≠ 0.
The relevant regression output needed to conduct the test of joint significance
appears in Table 15.6.
Total 29 0.1338
SYNOPSIS OF I N T R O D U C T O R Y CASE
EXERCISES 15.1
c. What is the conclusion to the test? Coefficients Error fS ta t p-value 95% 95%
Intercept 152.27 119.70 1.27 0.2142 − 93.34 397.87
3. In a simple linear regression based on 30 observations, it is
found that b1 = 7.2 and Consider the hypotheses: X1 12.91 2.68 4.81 5.06E-05 7.40 18.41
a. Calculate the value of the appropriate test statistic. a. Specify the hypotheses to determine whether x, is
b. At the 5% significance level, what is the critical value(s)? linearly related to y . At the 5% significance level, use
c. What is th e conclusion to the test? the p -value approach to com plete the test. Arex, and y
linearly related?
4. Consider the following regression results based on 20
observations. b. What is the 95% confidence interval for β 2? Using this
Standard
confidence interval, is x2 significant in explaining y ? Explain.
Lower Upper
Error t Stat p -value 95% 95% c. At the 5% significance level, can you conclude that β 1 is
34.2123 less than 20? Show the relevant steps of the appropriate
Intercept 4.5665 7.4920 0.0000 24.62 43.81
x1 0.1223 hypothesis test.
0.1794 0.6817 0.5041 − 0.25 0.50
7. The following ANOVA table was obtained w hen estimating a
a. Specify the hypothese s to determine if the intercept multiple regression model.
differs from zero. Perform this test at the 5% significance
ANOVA df SS MS F Significance F
level.
Regression 2 22016.75 11008.38 0.0228
b. Construct the 95% confidence interval for the slope
Residual 17 39286.93 2310.996
coefficient. At the 5% significance level, d oes the slope
Total 19 61303.68
differ from zero? Explain.
a. How many explanatory variables were specified in the a. Specify the estimated regression equation.
model? How many observations were used? b. At the 5% significance level, show whether
b. Specify the hypotheses to determine w hether the the poverty rate and the crime rate are linearly
explanatory variables are jointly significant. related.
c. Compute the value of th e test statistic. c. Construct a 95% confidence interval for the slope
d. At the 5% significant level, what is the conclusion to the coefficient of income. Using the confidence interval,
test? Explain. determine whether incom e is significant in explaining
the crime rate at the 5% significance level.
Applications d. At the 5% significance level, are the poverty rate
8. A recent study on the evolution o f mankind show s that, with and income jointly significant in explaining the
a few exceptions, world-record holders in the 100-meter crime rate?
dash have progressively gotten bigger over time (The W all 10. A model relating the return on a firm's stock as a function
Street Journal, July 22,2 0 09 ). The following table shows o f its price-to-earnings ratio and its price-to-sales ratio is
runners w h o have held the record, along with their record- estimated: Return = β 0 + β 1 P/E + β 2P /S + ε . A portion of
holding times and heights: the regression results follows.
Asafa Powell (2007) 9.74 75 Intercept − 12.0243 7.886858 − 1.5246 0.1390 − 28.21 4.16
Usain Bolt (2008) 9.69 77 P/E 0.1459 0.4322 0.3376 0.7383 − 0.74 1.03
Height − 0.0477 0.0163 0.0332 − 0.09 − 0.01 c. Are both explanatory variables individually significant at
the 10% significance level? Show the relevant steps of
a. Formulate the estimated regression equation. the appropriate hypothesis tests.
b. Specify the hypotheses to determine whether Height is 11. Akiko Hamaguchi is a manager at a small sushi restaurant in
linearly related to Time. Phoenix, Arizona. Akiko is concerned that the weak economic
c. Calculate the value of the appropriate test statistic. environment has hampered foot traffic in her area, thus causing
d. At the 5% significance level, is Height significant? a dramatic decline in sales. In order to offset the decline in sales,
Explain. she has pursued a strong advertising campaign. She believes
9. For a sam ple o f 20 New England cities, a sociologist studies advertising expenditures have a positive influence on sales. To
the crime rate in each city (crimes per 100,000 residents) as a support her claim, Akiko estimates a regression model as
function o f its poverty rate (in %) and its median incom e (in Sales =β 0 + β 1 Advertising + β 2 Unemployment + ε.
$1,000s). A portion o f the regression results are as follows. A portion of the regression results follows.
ANOVA df SS MS F Significance F
ANOVA df SS MS F Significance F
Intercept − 301.62 549.7135 -0 .5 4 8 7 0.5903 − 1,461.52 858.28 Intercept 17.5060 3.9817 4.3966 0.0006 8.97 26.05
Poverty 53.1597 14.2198 3.7384 0.0016 23.16 83.16 Unemployment − 0.6879 0.2997 − 2.2955 0.0377 − 1.33 − 0.05
Income 4.9472 8.2566 0.5992 0.5569 − 12.47 22.37 Advertising 0.0266 0.0068 3.9322 0.0015 0.01 0.04
a. A t th e 5% significance level, te st w h e th e r th e Reserve's website. A portion o f the regression results are
e xp la n a to ry variables jo in tly in flu e n ce sales. shown below (n = 60):
b. A t th e 1% sig n ifica n ce level, te st w h e th e r th e Standard Lower Upper
u n e m p lo y m e n t rate is n e g a tive ly related w ith sales. Coefficients Error t Stat p -value 95% 95%
c. A t th e 1 % sig n ifica n ce level, te s t w h e th e r a d ve rtisin g Intercept − 0.0038 0.0088 − 0.4273 0.6708 - 0.02 0.01
e xp e nd itu re s are p o s itiv e ly related w ith sales. D short 0.0473 0.0168 2.8125 0.0067 0.01 0.08
12. F I L E A re a lto r exam ines th e fa cto rs th a t in flu e n ce th e price
Use a 5% significance level in order to determine
o f a house in A rlin g to n , Massachusetts. He collects data on
whether there is a linear relationship betw een Dshort
recent house sales (Price) and notes each house's square
and Dlong.
fo o ta g e (Sqft) as w e ll as its n u m b e r o f b e droom s (Beds)
and n u m b e r o f b a thro o m s (Baths). A p o rtio n o f th e data 14. F IL E Caterpillar, Inc. manufactures and sells heavy
is show n. The e n tire data set is fo u n d on th e te x t w ebsite, construction equipment worldwide. The performance of
labeled A rlington Homes. Caterpillar's stock is likely to be strongly influenced by the
Price
economy. For instance, during the subprime mortgage crisis,
Sqft Beds Baths
the value of Caterpillar's stock plunged dramatically. Monthly
840000 2768 4 3.5
data for Caterpillar's risk-adjusted return and the risk-
822000 2500 4 2.5
adjusted market return are collected for a five-year period
(n = 60). A portion o f the data is shown below. The entire
307500 850 1 1 data set is found on the text website, labeled Caterpillar.
Date
R − Rf
a. Estimate: Price = β 0 + β1 Sqft + β 2Beds + β 3Baths + ε. R M − R f
Show the regression results in a well-formatted table. 1/1/2006 17.66 2.21
Dlong = β0 + β1 Dshort + ε , where Dlong is the change C a te rp illa r is riskier th a n th e m arke t (beta sig n ific a n tly
in the long-term interest rate (10-year Treasury bill) and gre a te r th a n 1).
Dshort is the change in the short-term interest rate (3-month c. A t th e 5% sig n ifica n ce level, is th e re evidence o f a b n orm al
Treasury bill). Monthly data from January 2006 through returns?
December 2010 were obtained from the St. Louis Federal
When testing linear restrictions, the test statistic is assumed to follow the
distribution with d f1equal to the number of linear restrictions and df2 = n − k − 1,
where k is the number of explanatory variables in the unrestricted model. Its value
is calculated as
where SSER and SSEU are the error sum of squares of the restricted and the unre
stricted models, respectively.
EXAMPLE 15.5
A manager at a car wash company in Missouri wants to measure the effectiveness
of price discounts and various types of advertisement expenditures on sales. For
the analysis, he uses varying price discounts (Discount) and advertisement expen
ditures on radio (Radio) and newspapers (Newspaper) in 40 counties in Missouri.
A portion of the monthly data on sales (in $ 1,000s), price discounts (in percent),
and advertisement expenditures (in $ 1,000s) on radio and newspapers are shown in
Table 15.8; the complete data are available on the text website, labeled Car wash. At
the 5% level, determine if the advertisement expenditures on radio and newspaper
have a significant influence on sales.
We will use the critical value approach to conduct the test. We use d f1 = 2, since we
are testing for two restrictions, β 2 = 0 and β 3 = 0, and df2 = n − k − 1 = 4 0 − 3 −
1 = 36. Taking the appropriate SSE values from Table 15.9, we calculate the value
of the relevant test statistic as
With a = 0.05, d f1 = 2, and df2 = 36, we find the approximate critical value as
F 0.05,(2,36) = 3.26, or equivalently, we use Excel’s function F.INV.RT(0.05,2,36)
and obtain 3.26. We reject H0 since 14.52 is greater than 3.26. At the 5% level,
we conclude that the advertisement expenditures on radio and newspaper have a
significant influence on sales.
H 0: β 2
= β3
H A: β 2 ≠ β 3
We again specify the unrestricted model (U) as
(U) Sales = β 0 + β1 Discount + β 2Radio + β3,Newspaper + ε .
In order to implement the partial F test, we then create the restricted (R) model.
Note that under the restriction that β2 = β 3, the unrestricted model simplifies to
Sales = β 0 + β1 Discount + β 2Radio + β 2Newspaper + ε , that is,
N otes : Parameter estimates are in the main body o f the table w ith the p-values in
parentheses; * represents significance at 5% level. The last row presents the error sum of
squares.
We will use the p -value approach to conduct the test. We use d f 1= 1 since we are
testing for only one restriction, β2 = β 3, and df2 = n − k − 1 = 40 − 3 − 1 = 36.
Using the appropriate SSE values from Table 15. 10, we calculate the value of the
test statistic as
With d f1 = 1 and df2 = 36, we use Excel to compute the p -value as 0.2072 (EDIST.
RT( 1.65,1,36) = 0.2072). We do not reject H0 since the p -value > α = 0.05. At the
5% significance level, we cannot conclude that the influence of the advertisement
expenditures on radio is different from the influence of the advertisement expendi
tures on newspapers.
EXERCISES 15.2
While such predictions can be useful, they are subject to sampling variations. In other
words, the predicted value is a point estimate that ignores sampling error. In this section
we focus on the interval estimates for a prediction from the OLS regression equation.
As in the case of tests of significance, we assume that the OLS estimators are normally
distributed.
We will construct two kinds of interval estimates regarding y:
1. A confidence interval for the expected value of y.
2. A prediction interval for an individual value of y.
It is common to call the first one a confidence interval and the second one a prediction
interval. Consider the following multiple regression equation:
where x10, x20, . . . ,x k0denote specific values fo r x1, x2,...,xk at w hich y0 is evaluated. In the
above example, we used x 10= 0.25 and x 20= 4. Alternatively, we can evaluate the expected
value of the response variable at x10, x20, ... ,xk0 as
The expected value equation uses the fact that the expected value of the error term is
assumed to be zero, that is, E(ε0) = 0; we discuss this assumption in the next section.
Note that the prediction interval will be wider than the confidence interval because it also
incorporates the error term ε0 . Intuitively, it is easier to predict the average value for a
response variable as compared to its individual value. We first derive a confidence interval
for E( y0), followed by a prediction interval for y0.
The predicted value, ŷ0 = b0 + b 1x1 + b 2x 20 + . . . + bkxk0, is the point estimate
for E(y0). In our earlier example, 0.47 is the point estimate of E (y0) when x 10 = 0.25 and
x 20= 4. We form a 100(1 − a )% confidence interval for E(y0) as ŷ0 ± tα / 2,dfSe( ŷ0 ), where
se( ŷ0 ) is the standard error of the prediction. While there is a simple formula to compute
the standard error se( ŷ0 ) for a simple regression model, it is very cumbersome to do so for
a multiple regression model. We describe a relatively easy way to construct a confidence
interval that works for both simple and multiple regression models.
For specific values of x 1, x2, . . . , xk, denoted by x10, x20 . . . , xk0, a 100( 1 − a ) % con
fidence interval of the expected value of y is computed as
ŷ0 ± tα/ 2,dfSe( ŷ0) ,
where ŷ0 = b0 + b 1x1 + b2x 2 + . . . + bkxk0,se( ŷ0 ) is the standard error o f ŷ0, and
d f = n − k − 1.
To derive ŷ0 and se( ŷ0 ) we first estimate a modified regression model where y
is the response variable and the explanatory variables are defined as x* = x 1 − x10,
x 2*= x 2 − x20
, . . . , x k*= xk − xk0.The resulting estimate of the intercept and its stan
dard error equal ŷ0 and se( ŷ0 ), respectively.
EXAMPLE 15.7
We again reference the data from Table 15.1 and the regression model Win = β 0 +
β 1 BA + β 2ERA + ε . Construct a 95% confidence interval for expected winning
percentage if BA is 0.25 and ERA is 4.
SOLUTION: Let y, x 1, and x2 denote Win, BA, and ERA, respectively. In order
to construct a confidence interval for E( ŷ0 ), we follow the above-mentioned
procedure to derive ŷ0 as well as se(ŷ0 ). First, we define two modified explanatory
variables as x1* = x 1 − 0.25 and x2*= x2 − 4. Table 15.11 shows the computations
of their values.
TABLE 15.11 Com puting th e Values of Modified Explanatory Variables (Example 15.7)
The regression output with y as the response variable and x 1*and x 2*as the explana
tory variables is presented in Table 15.12.
TABLE 15.12 Regression Results with Modified Explanatory Variables (Example 15.7)
Regression Statistics
Multiple R 0.8459
R Square 0.7156
Adjusted R
Square 0.6945
Standard
Error 0.0375
Observations 30
ANOVA
df SS MS F Significance F
Regression 2 0.0958 0.0479 33.9663 4.25E-08
Residual 27 0.0381 0.0014
Total 29 0.1338
Standard Lower Upper
Coefficients Error t Stat p -value 95% 95%
Intercept 0.4847 0.0085 57.2582 0.0000 0.4673 0.5021
The modified regression output is identical to the original regression output (see
the summarized results for Model 3 in Table 15.7) except for the estimates of the
intercept term. The boldface intercept estimate is 0.4847 and its standard error is
0.0085. Therefore, we use ŷ0 = 0.4847 and se(ŷ0 ) = 0.0085 in constructing the con
fidence interval. Note that Excel’s calculation fo r ŷ0 is slightly different from our
earlier estimate, ŷ0 = 0.13 + 3.28 × 0.25 − 0.12 × 4 = 0.47; this is simply due to
rounding.
For a 95% confidence level and d f = n − k − 1 = 30 − 2 − 1 = 27, we find
ta/2,df = t0.025,27 = 2.052. The 95% confidence interval for E(ŷ0 ) is
ŷ0 ± ta/2,df se(ŷ0 ) = 0.4847 ± 2.052 × 0.0085 = 0.4847 ± 0.0174.
Or, with 95% confidence,
0.4673 ≤ E(ŷ0 ) ≤ 0.5021.
Given BA of 0.25 and ERA of 4, we have 95% confidence that the average winning
percentage lies between 0.4673 and 0.5021. Note that these calculations are also
provided in the boldface Lower 95% and Upper 95% values of Table 15.12.
As mentioned earlier, the prediction interval pertains to the individual value of the
response variable defined for specific explanatory variables as y0 = β 0 + β1 x10 + β2x20 +
. . . + β kxk0+ ε0 . The prediction interval is wider than the confidence interval because it
incorporates the variability of the random error term ε0 .
P R E D IC T IO N IN T ER VA L FOR AN I N D I V I D U A L VALUE OF y
where d f = n − k − 1, se(ŷ0) is the standard error o f ŷ0, and se is the standard error
of the estimate.
EXAMPLE 15.8
Reconsider the estimated m odel, ŷ0 = 0.13 + 3.28x1 − 0.12 x 2, where y, x 1, and x 2
denote Win, BA, and ERA, respectively.
a. Construct the 95% prediction interval for Win if BA is 0.25 and ERA is 4.
b. Comment on any differences between this prediction interval and the confi
dence interval constructed in Example 15.7.
S O L U T IO N :
Concepts
24. In a simple linear regression based on 30 observations, the 3 4 6 2 5 4 8 0
following information is provided: ŷ = −6.92 + 1.35x and
40 53 80 42 70 50 110 38
se = 2.78. Also, se(ŷ0 ) evaluated a t x = 30 is 1.02.
a. Construct a 95% confidence interval for E(y ) if x = 30. a. Construct a 90% confidence interval fo r the expected salary
fo r an individual w h o com pleted 6 years o f higher education.
b. Construct a 95% prediction interval fory if x equals 30.
b. C o n stru ct a 90% p re d ic tio n in te rva l fo r salary fo r an
c. Which interval is narrower? Explain.
in d iv id u a l w h o co m p le te d 6 years o f h ig h e r e d u cation.
25. In a multiple regression with 40 observations, the following
c. C om m ent on th e difference in th e w id th s o f these intervals.
sample regression equation is obtained: ŷ = 12.8 + 2.6x 1 −
1 .2x2 with se = 5.84. Also, w h en x 1 equals 15 and x2 equals 6, 29. W ith th e data in th e accom panying ta b le , estim ate
Recall from Chapter 7 that desirable properties of an estimator include unbiasedness and
efficiency. In the regression context, an estimator bj is unbiased if E(bj) = β j, or in other
words, the average value of the estimator equals the unknown population parameter.
Efficiency is implied when an unbiased estimator also has the minimum variance among
the class of all unbiased estimators; that is, the estimator estimates the parameter of inter
est in the best possible manner.
Under the assumptions of the classical linear regression model, OLS provides the
best linear unbiased estimator (BLUE) in that it is unbiased and efficient. These desirable
properties of the OLS estimator become compromised as one or more model assumptions
are violated. Aside from parameter estimates, the validity of the significance tests is also
influenced by the assumptions. For certain violations, the standard errors of the estimators
are biased; in these cases it is not possible to make meaningful inferences from the t and
the F test results.
The assumptions of the classical linear regression model are, for the most part, based
on the error term ε . Since the residuals, or the observed error term, e = y − ŷ , contain
useful information regarding ε , it is common to use the residuals to investigate the as
sumptions. In this section, we will rely on residual plots to detect common violations
to the assumptions. These graphical plots are easy to use and provide informal analysis
of the estimated regression models. Formal tests are beyond the scope of this book.
R E S I D U A L PLOTS
It is common to plot the residuals e on the vertical axis and the explanatory variable xj
on the horizontal axis. Such plots are useful for detecting deviations from linearity, dis
cussed in the next chapter, as well as heteroskedasticity. For a multiple regression model,
we can also plot the residuals against the predicted value ŷ of the model. Finally, if the
regression is based on time series data, we can plot the residuals against time to detect
serial correlation.
In Figure 15.1, we present a residual plot when none of the assumptions has been vio
lated. (Excel computes the residuals and also plots them against all explanatory variables.
After choosing Data > Data Analysis > Regression, we select Residuals and Residual
Plots in the regression dialog box.)
Note that all the points are randomly dispersed around the zero value of the residuals.
As we will see below, any discernible pattern of the residuals indicates that one or
more assumptions have been violated.
Detection
The detection methods for multicollinearity are mostly informal. The presence of a high
R2 coupled with individually insignificant explanatory variables can be indicative of mul
ticollinearity. Sometimes researchers examine the correlations between the explanatory
variables to detect severe multicollinearity. One such guideline suggests that multicol
linearity is severe if the sample correlation coefficient between any two explanatory vari
ables is more than 0.80 or less than − 0.80. Seemingly wrong signs of the estimated
regression coefficients may also be indicative of multicollinearity.
EXAMPLE 1 5 .9 F IL E
Revisit the introductory case, where we used the data labeled Baseball to estimate
a multiple regression model as Examine the
multicollinearity issue in this model.
S O L U T IO N : We select the data on BA and ERA, and use the CORREL function
in Excel to compute the sample correlation rBA,ERA = 0.0581. The fact that the
sample correlation is very weak is not surprising, since we do not expect a team ’s
batting average to be correlated with its earned run average. After all, a team ’s
pitching staff determines ERA, whereas the remaining players on the team
predominantly influence BA. We conclude that multicollinearity is of no concern
in this application.
Remedy
Inexperienced researchers tend to include too many explanatory variables in their
quest not to omit anything important and in doing so may include redundant vari
ables that essentially measure the same thing. When confronted with m ulticollinear
ity, a good remedy is to drop one of the collinear variables if we can justify its
redundancy. Another option is to obtain more data, since the sample correlation
may get weaker as we include more observations. Sometimes it helps to go back
to the drawing board and reformulate a model with transform ed variables that may
not be collinear. At times, the best approach may be to do nothing, especially if the
estimated model yields a high R 2, which implies that the estim ated model is good
for prediction as is.
Detection
We can use informal residual plots to gauge heteroskedasticity. The residuals of the esti
mated model are plotted against the explanatory variables or against the predicted values y.
There is no heteroskedasticity if the residuals are randomly dispersed across the values of
xj o r ŷ. On the other hand, heteroskedasticity exists if the variability increases or decreases
over the values of xj or ŷ.
EXAMPLE 1 5 .1 0
Consider a simple regression model that relates monthly sales (Sales) from a chain
of convenience stores with the square footage (Sqft) of the store. A portion of
the data used for the analysis is shown in Table 15.13; the complete data, labeled
C onvenience Stores, are available on the text website. Estimate the model and use
a residual plot to determine if there is heteroskedasticity in the data.
110 1470
Note that the residuals seem to fan out across the horizontal axis. Therefore, we
conclude that heteroskedasticity is a likely problem in our application relating sales
to square footage. This result is not surprising, since you would expect sales to vary
more as square footage increases. For instance, a small convenience store is likely
to include only bare essentials for which there is a fairly stable demand. A larger
store, on the other hand, may include specialty items resulting in more fluctuation
in sales.
Remedy
As mentioned earlier, there are two problems associated with heteroskedasticity. First,
while the estimators are unbiased, they are not efficient. Second, even if we ignore the
issue of efficiency, we cannot conduct an important aspect of statistical analysis, namely
the tests of significance. The t and F tests are no longer valid, since the standard errors
of the estimators are biased. This has prompted some researchers to use the OLS esti
mates along with a correction for the standard errors, called W hite’s correction. Many
statistical computer packages routinely make this correction, thus enabling researchers
to perform legitimate t and F tests. Unfortunately, the current version of Excel does not
have the ability to make this correction. Other methods for correction are outside the
scope of this book.
Common Violation 3: The Error Term Is Serially Correlated
When obtaining the OLS estimators, we assume that the error term is uncorrelated
across observations; that is, there is no serial correlation. This assumption often
breaks down in studies with time series data where the errors are likely to exhibit
positive serial correlation. Variables such as GDP, employment, and asset returns ex
hibit business cycles. As a consequence, successive observations are not independent
o f each other. The omission o f important explanatory variables can also cause serial
correlation.
In the presence of serial correlation, the OLS estimators are generally unbiased,
though not efficient. In other words, they are no longer BLUE. In addition, with
positive serial correlation (the most common form o f serial correlation), the standard
errors are biased downward, making the model look better than it really is with a
spuriously high R2. Furthermore, the inflated values of the tdf and statistics may
suggest that the explanatory variables are individually and jo in tly significant when
this is not true.
Detection
We can plot the residuals against time to look for serial correlation. I f the residuals show
no pattern around the horizontal axis, then serial correlation is not likely a problem. Posi
tive serial correlation is implied if a positive residual in one period is followed by positive
residuals in the next few periods, followed by negative residuals for a few periods, then
positive residuals, and so on. Regression models estimated using time series data often
suffer from positive serial correlation. On the other hand, negative serial correlation is
implied if a positive residual is followed by a negative residual, then a positive residual,
and so on.
EXAMPLE 15.11
Consider y = β 0 + β1x1 + β 2x2 + ε where y represents sales at a sushi res
taurant and x 1 and x2 represent advertising costs and the unemployment rate,
respectively. A portion of monthly data from January 2008 to June 2009 is given
in Table 15.14; the complete data, labeled Sushi Restaurant, are on the text
website. Inspect the behavior of the residuals in order to comment on serial
correlation.
TABLE 1 5 . .14 Sales, Advertising Costs, and Unem ploym ent Data for Example 15.11
Remedy
Since the O LS estim ators are unbiased under the presence o f serial co rrela tion ,
researchers often use the O LS estim ates but correct the standard errors using the
N ew ey-W est procedure. As in the case o f heteroskedasticity, m any sta tistical com
puter packages have the capacity to make th is co rrectio n; u nfo rtun ately, the current
version o f E xcel does not have this c a p a b ility. We can p erform le g itim a te t and F tests
once the standard errors have been corrected. O ther m ethods fo r co rrectio n are outside
the scope o f th is book.
Remedy
It is comm on to use the instrum ental variable technique to address the endogeneity
problem . We fin d an instrum ental variable that is h ig h ly correlated w ith the endogenous
explanatory variable but not w ith the error term . Further discussion o f the instrum ental
variable approach is outside the scope o f this text.
EXERCISES 15.4
Concepts a. Access the data labeled Ann A rbo r R ental from the text
website and estim ate Rent = 1 0 + β1 Beds + β 2Baths +
32. Using 20 observations, the m ultiple regression model
β 3Sqft + ε .
y = β 0 + β 1 x1 + β2 x2 + ε was estim ated. Excel produced
b. W hich o f the explanatory variables m ight cause
the fo llo w in g relevant results.
heteroskedasticity in th e model? Explain.
df SS MS F Significance F
c. Use residual plots to verify your econom ic in tu itio n .
Regression 2 2.12E + 12 1 .06E + 12 56.5561 3.07E-08
37. F IL E H ealthy livin g has always been an im p o rta n t goal
Residual 17 3 .1 9 E + 11 1.88E + 10
fo r any society. In a recent ad cam paign fo r Walt Disney,
Total 19 2 .4 4 E + 12
First Lady M ichelle Obama shows parents and children
S ta n d a r d Low er U pper
th a t eating w ell and exercising can also be fu n (USA Today,
C oefficients Error t St a t p - v a lu e 95% 95%
Septem ber 3 0 , 2010). Consider a regression m odel
Intercept − 987557 131583 − 7.5052 0.0000 − 1265173 − 709941
th a t conjectures th a t fru its and vegetables and regular
X1 29233 32653 0.8952 0.3832 − 39660 98125
exercising have a positive e ffect on health and sm oking
X2 30283 32645 0.9276 0.3666 − 38592 99158 has a negative effect on health. The sam ple consists o f the
a. At the 5% significance level, are the explanatory percentage o f these variables observed in various states
a. Graph th e residuals e against the values o f the a. Estimate the model Healthy = β 0 + β 1 Fruits/Vegetables +
explanatory variable x and look fo r any discernible β 2Exercise + β3 Smoke + ε .
pattern. b. Analyze the data to determ ine if m ulticolline arity and
b. W hich assum ption is being violated? Discuss its heteroskedasticity are present.
consequences and suggest a possible remedy. 38. F ILE A capital asset pricing m odel (CAPM) for Johnson &
34. A sim ple linear regression, y = β0 + β1 x + ε, is estim ated Johnson (J&J) was discussed in Example 15.3. The m odel
w ith tim e series data.The resulting residuals e and the tim e uses the risk-adjusted stock return R − Rf fo r J&J as the
variable fa re shown below. response variable and the risk-adjusted m arket return
W R IT IN G W IT H STATISTICS
p r o fe s s io n a l lif e t h r o u g h 2 0 0 8 . F o r e x p o s it io n , B A h a s b e e n m u l t i p l i e d b y 1 0 0 0 .
In an attempt to assess the factors that influence an outfie lde r’s salary in M L B . data were Sample
collected from 138 current players. Management believes that an o utfie lde r’s salary is
best predicted using the o utfie lde r’s overall batting average (B A ), runs batted in (R B I), Report—
and years o f experience (Experience) as an M L B player. Table 15.A provides some de
scriptive statistics on these relevant variables. Baseball
The average salary o f an M L B outfielder in 2008 is a staggering $3,459,000; however,
the m inim um salary o f $390,000 and the m axim um salary o f $18,623,000 suggest quite Salaries
TABLE 15.A D escrip tiv e S ta tistic s o n Salary, BA, RBI, a n d E xp erien ce, n = 138
Salary
(in $1,000s) BA RBI Experience
Mean 3,459 271 43 6
M inimum 390 152 1 1
M aximum 18,623 331 102 20
a bit of variability in salary. The average outfielder has a BA of 271 with 43 RBIs in a
season. Experience of outfielders in 2008 varied from only 1 year to 20 years, with an
average of 6 years.
Table 15.B provides regression results from estimating a model where BA, RBI, and
Experience are the explanatory variables and Salary is the response variable.
TABLE 1 5 .B A nalysis o f Salary o f B aseball P layers
Variable C o efficien t
Intercept − 4769.40 (0.1301)
BA 4.76 (0.6984)
RBI 80.44* (0.0000)
Experience 539.67* (0.0000)
R2 = 0.58
F(3,133) = 61.54 (associated p-value = 0.0000)
N otes: p - v a l u e s a r e in p a r e n t h e s e s ; *d e n o t e s s i g n if ic a n c e a t t h e 5 % le v e l.
All slope coefficients are positive, suggesting a positive relationship between all three ex
planatory variables and Salary. For instance, the slope coefficient of Experience indicates
that if an outfielder stays in the major leagues for one additional year, then on average,
his salary will increase by $539,670, holding BA and RBI constant. The p -value associ
ated with the value of the F(3,133) test statistic shows that the explanatory variables are
jointly significant at the 5% level. Upon testing the explanatory variables individually, the
extremely small p-values associated with RBI and Experience reveal that these variables
have a significant linear relationship with Salary; surprisingly, BA is not significant at the
5% level. The coefficient of determination R 2 shows that 58% of Salary is explained by the
estimated regression model, leaving 42% of the variability in Salary unexplained.
Lastly, for an MLB player with seven years’ experience and an average BA and RBI of
266 and 50, respectively, the model predicts a salary of $4,295,320. With 95% confidence,
expected salary will lie between $3,731,360 and $4,859,280. Perhaps before management
makes an offer to the player, the model should consider including other factors that may
significantly influence a player’s salary. One possible explanatory variable for inclusion
is a player’s on-base percentage.
Conceptual Review
LO 15.1 C onduct te sts o f individual significance.
The A N O V A table from computer output provides both the value o f the test statistic
and its associated p -value.
la te d as w h e r e S S E R a n d S S E U a re th e e r r o r s u m o f s q u a re s
o f th e r e s t r ic t e d a n d th e u n r e s t r ic t e d m o d e ls , r e s p e c t iv e ly . I f th e n u l l h y p o th e s is is
r e je c t e d , w e c o n c lu d e t h a t th e l i n e a r r e s t r ic t io n s a re n o t v a lid .
t i o n i n t e r v a l f o r a n i n d i v i d u a l v a l u e o f y is g iv e n b y w h e re
df = n − k − 1, s e (ŷ 0 ) is th e s ta n d a r d e r r o r o f ŷ 0 , a n d se is th e s ta n d a r d e r r o r o f th e
e s tim a te .
Estimate: SAT = β 0 + β1 Incom e + β 2GPA + ε . Find th e ir sam ple co rre la tio n coefficients to co nfirm .
b. A t th e 5% significance level, are incom e and GPA jo in tly Generally, a h ig h P /E ra tio suggests th a t investors are
significant? Show th e relevant steps o f th e hypothesis test, e x p e ctin g h ig h e r earnings g ro w th in th e fu tu re com pared
using th e critical value approach. to com panies w ith a lo w e r P/E ratio. Investors use th e P/S
ra tio to d e te rm in e h o w m uch th e y are p a ying fo r a d o lla r o f
c. Predict SAT if Incom e is $80,000 and GPA is 3.5.
th e firm 's sales ra th e r th a n a d o lla r o f its earnings (P /E ratio).
Use these values fo r th e e x p la n a to ry variables to
In sh o rt, th e h ig h e r th e P/E ra tio and th e lo w e r th e P/S
c o n s tru c t a 95% p re d ic tio n in te rva l fo r th e in d ivid u a l
SAT score. ratio, th e m ore a ttra ctiv e th e in ve stm e n t. The ta b le b e lo w
shows th e 2010 an n ua l returns, th e P /E ratios, and th e P/S
44. F I L E G eorge believes th a t retu rn s o f m u tu a l fu n d s are
ratios fo r a p o rtio n o f th e 30 firm s in c lu d e d in th e D ow Jones
in flu e n ce d by a n n ua l tu rn o v e r rates and a n n ua l expense
Ind ustria l Average. The co m p le te data set can be fo u n d on
ratios. In o rd e r to s u b sta n tia te his claim , he ra n d o m ly
th e te x t w ebsite, labeled D o w 2 0 1 0 .
selects e ig h t m u tu a l fu n d s and co lle cts data on each fund's
fiv e -y e a r an n u a l re tu rn (Return), its an n ua l h o ld in g tu rn o v e r DOW
rate (Turnover), and its an n ua l expense ra tio (Expense). The Components Return (in %) P/E ratio P/S ratio
d ata set can be fo u n d on th e te x t w e b site , labeled Turnover 3M Co. 4.4 14.37 2.41
a n d Expense. Alcoa Inc. − 4.5 11.01 0.78
Artisan SOURCE: T h e 2 0 1 0 r e tu r n s ( J a n u a r y 1 , 2 0 1 0 – D e c e m b e r 3 1 , 2 0 1 0 ) w e r e o b t a in e d f r o m T he
in flu e nce a city's crim e rate. For 41 cities, she collects th e 47. F I L E A n u tritio n is t w a n ts to u n derstand th e influence o f
crim e rate (crimes per 100,000 residents), th e p o v e rty rate incom e and he a lthy fo o d on th e incidence o f sm oking. He
(in %), th e m edian in co m e (in $ 1,000s), th e p ercent o f collects 2009 data on th e percentage o f smokers in each
residents yo u n g e r th a n 18, and th e p e rcent o f residents o ld e r state in th e U.S. and th e co rresponding m edian incom e
th a n 65. A p o rtio n o f th e data is show n; th e e n tire data set and th e percentage o f th e p o p u la tio n th a t regularly eats
can be o b ta in e d fro m th e te x t website, labeled Crime. fru its and vegetables. A p o rtio n o f th e data is show n in th e
a ccom panying ta ble; th e co m p le te data set can be fo u n d on sh o w n in th e acco m p a n yin g table; th e co m p le te data set can
th e te x t w ebsite, labeled Smoking. _______________ be fo u n d on th e te x t w ebsite, labeled MCAS.
Fruits/Vegetables Median TSAL INC
State Smoke (%) (%) income Score STR (%) (in $1,000s) (in $1,000s) SGL (%)
AK 14.6 23.3 61,604 227.00 19.00 44.01 48.89 4.70
AL 16.4 20.3 39,980 230.67 17.90 40.17 43.91 4.60
a. Estim ate: Sm oke = β0+β1 F ruits/Vegetables + a. Estim ate Score = β 0 + β 1 STR + β 2TSAL + β 3INC +
β 2M edian Incom e + ε . β 4SGL + ε and sh o w th e regression results in a w e ll-
b. A t th e 5% level o f significance, are th e e xp la na to ry fo rm a tte d table.
variables in d iv id u a lly and jo in tly significant? Explain. b. Suppose you w a n t to te s t if school in p u t factors, STR
c. Use th e sam ple c o rre la tio n co e fficie nts to evaluate th e and TSAL, are s ig n ific a n t in e xp la in in g Score. Specify th e
p o te n tia l p ro b le m o f m u ltic o llin e a rity. c o m p e tin g hypotheses. Estim ate th e restricted m odel. At
48. F IL E A researcher exam ines th e fa ctors th a t in fluence th e 5% significance level, can you co n clu de th a t STR and
s tu d e n t perform ance. She gathers data on 224 school TSAL are jo in tly significant?
districts in Massachusetts. The response variable is th e c. Suppose you w a n t to te s t if so cio e con o m ic factors, INC
stu d e n ts 'm e a n score on a standardized te st (Score). She and SGL, are s ig n ific a n t in e xp la in in g Score. Specify th e
uses fo u r e x p la n a to ry variables in her analysis: th e stu d e n t- c o m p e tin g hypotheses. Estim ate th e restricted m odel. At
to -te a ch e r ratio (STR), th e average teacher's salary (TSAL), th e 5% significance level, can you c o n clu de th a t INC and
th e m edian household incom e (INC), and th e percentage SGL are jo in tly significant?
o f single fa m ily households (SGL). A p o rtio n o f th e data is
CASE STUDIES
C a s e S tu d y 15.1
A m e r ic a n f o o t b a l l is th e h ig h e s t - p a y in g s p o r t o n a p e r - g a m e b a s is . G iv e n th a t th e q u a r te r
b a c k is c o n s id e r e d th e m o s t im p o r t a n t p la y e r o n th e te a m , h e is t y p i c a ll y w e ll- c o m p e n s a te d .
A s p o r ts s t a t is t ic ia n e x a m in e s th e fa c t o r s th a t in f lu e n c e a q u a r t e r b a c k ’ s s a la r y ( S a la r y ) . H e
b e lie v e s th a t a q u a r t e r b a c k ’ s p a s s c o m p le t io n ra te ( P C ) is th e m o s t im p o r t a n t v a r ia b le
a f f e c t in g S a la r y . T h e s t a t is t ic ia n a ls o w o n d e r s h o w t o t a l to u c h d o w n s s c o r e d ( T D ) a n d a
q u a r t e r b a c k ’ s a g e ( A g e ) m ig h t im p a c t S a la r y . T h e s t a t is t ic ia n c o lle c t s 2 0 0 9 d a ta o n S a la r y ,
P C , T D , a n d A g e . A p o r t io n o f th e d a ta is s h o w n i n th e a c c o m p a n y in g t a b le ; th e c o m p le te
d a ta s e t c a n b e f o u n d o n th e t e x t w e b s ite , la b e le d Q u a r t e r b a c k S a l a r i e s .
D a t a f o r C a s e S t u d y 1 5 . 1 Q u a rte r b a c k S a la ry D a ta , 2 0 0 9
SOURCE: U S A T o d a y d a ta b a s e f o r s a la rie s ; N F L .c o m f o r o t h e r d a ta .
I n a r e p o r t, u s e th e s a m p le in f o r m a t io n to :
1. E s tim a te a n d in te r p r e t th e m o d e l: S a la r y = β 0 + β 1 P C + β 2 T D + β 3 A g e + ε .
2. D is c u s s th e in d iv id u a l a n d jo in t s ig n ific a n c e o f th e e x p la n a to ry v a ria b le s a t th e 5 % le v e l.
3. D e te r m in e w h e th e r T D a n d A g e a re j o i n t l y s ig n if ic a n t a t th e 5 % s ig n ific a n c e le v e l.
4. C o n s tr u c t a 9 5 % c o n fid e n c e in t e r v a l f o r th e e x p e c te d s a la ry o f a q u a rte r b a c k w it h
a v e ra g e v a lu e s o f P C T , T D , a n d A g e .
Case Study 15.2
A p p le In c . h a s e s ta b lis h e d a u n iq u e r e p u ta tio n in th e c o n s u m e r e le c tr o n ic s in d u s tr y w it h
its d e v e lo p m e n t o f p ro d u c ts s u c h as th e iP o d , th e iP h o n e , a n d th e iP a d . A s o f M a y 2 0 1 0 ,
A p p le h a d s u rp a s s e d M ic r o s o f t as th e m o s t v a lu a b le c o m p a n y in th e w o r ld ( T h e N e w Y o r k
T im e s , M a y 2 6 , 2 0 1 0 ) . M ic h a e l G o m e z is a s to c k a n a ly s t a n d w o n d e rs i f th e r e tu r n o n
A p p le ’ s s to c k is b e s t m o d e le d u s in g th e C A P M m o d e l. H e c o lle c ts fiv e y e a rs o f m o n th ly
d a ta , a p o r t io n o f w h ic h is s h o w n . T h e f u l l d a ta s e t c a n b e fo u n d o n th e te x t w e b s ite ,
la b e le d A p p l e .
D a ta f o r C a s e S tu d y 1 5 .2 A p p le R e tu rn D a ta , n = 60
F IL E
SOURCE: F in a n c e .y a h o o .c o m a n d U.S. T re a s u ry .
I n a r e p o r t, u se th e s a m p le in f o r m a t io n to :
1. E s tim a te a n d in te r p r e t C A P M : R − R f = β 0 + β 1 ( R M − R f) + ε . S e a rc h f o r A p p le ’ s
r e p o r te d B e ta o n th e W e b a n d c o m p a re i t w it h y o u r e s tim a te .
2. A t th e 5 % s ig n ific a n c e le v e l, is th e s to c k r e tu r n o f A p p le r is k ie r th a n th a t o f th e
m a rk e t? A t th e 5 % s ig n ific a n c e le v e l, d o a b n o r m a l re tu rn s e x is t? E x p la in .
Case S tu d y 15.3
A c c o r d in g to a r e c e n t r e p o r t b y th e g o v e rn m e n t, n e w h o m e c o n s tr u c tio n f e ll to a n
1 8 -m o n th lo w in O c to b e r, 2 0 1 0 ( C N N M o n e y .c o m , N o v e m b e r 1 7, 2 0 1 0 ) . H o u s in g s ta rts ,
o r th e n u m b e r o f n e w h o m e s b e in g b u ilt , e x p e rie n c e d a n 1 1 .7 % d ro p in its s e a s o n a lly
a d ju s te d a n n u a l ra te . U r m il S in g h w o r k s f o r a m o rtg a g e c o m p a n y in M a d is o n , W is c o n s in .
S h e w a n ts to b e tte r u n d e rs ta n d th e q u a n tita tiv e r e la tio n s h ip b e tw e e n h o u s in g s ta rts , th e
m o r tg a g e ra te , a n d th e u n e m p lo y m e n t ra te . S h e g a th e rs s e a s o n a lly a d ju s te d m o n th ly d a ta
o n th e s e v a ria b le s f r o m 2 0 0 6 :0 1 − 2 0 1 0 :1 2 . A p o r t io n o f th e d a ta is s h o w n b e lo w ; th e c o m
p le te d a ta , la b e le d H o u s i n g S t a r t s , a re a v a ila b le o n th e te x t w e b s ite .
D a ta f o r C a s e S tu d y 1 5 .3 H o u s in g S ta rts a n d O th e r F a ctors, n = 6 0
F IL E
Housing Starts Mortgage Unemployment
Date (in 1,000s) Rate (%) Rate (%)
2006–01 2273 6.15 4.7
2006–02 2119 6.25 4.8
SOURCE: C e n s u s B u re a u a n d B o a r d o f G o v e r n o rs .
I n a r e p o r t, u se th e s a m p le in f o r m a t io n to :
2. A t th e 5 % s ig n ific a n c e le v e l, e v a lu a te th e in d iv id u a l a n d j o in t s ig n ific a n c e o f th e
e x p la n a to r y v a ria b le s .
be able to:
LO 1 6 .3 D e s c rib e t h e m e th o d u sed t o c o m p a r e lin e a r w it h lo g tr a n s fo r m e d m o d e ls .
Regression analysis is one o f the most w idely used statistical techniques in business, engineering, and
the social sciences. It empirically validates not only w h ether a relationship exists between variables,
but also quantifies the strength o f the relationship. So far, we have considered only linear regression
models, w h ether single- or multiple-variable. There are numerous applications where the relation
ship between the explanatory variable and the response variable cannot be represented by a straight
line and, therefore, must be captured by an appropriate curve. In fact, the choice o f a functional form
is a crucial part o f specifying a regression model. In this chapter, we discuss some com m on nonlinear
regression models by making simple transformations o f the variables. These transformations include
squares and natural logarithms, which capture interesting nonlinear relationships w hile still allowing
easy estimation w ith in the fram ework o f a linear regression model. We will use numerical measures
to choose between alternative model specifications.
I N T R O D U C T O R Y C A S E
Source: w w w .z illo w .c o m .
THE Q U A D R A T IC REGRESSION M O D E L
9.62 11 121
SOLUTION:
a . It is always informative to begin with a scatterplot of the response variable
against the explanatory variable. Make sure that the response variable is on the
vertical axis. Figure 16.2 shows average cost against output. We also superim
pose linear and quadratic trends on the scatterplot (in Excel, right click on the
scatterpoints, add Trendline, and choose Linear and Polynomial with Order 2).
At lower and higher levels of output, average costs are highest. It appears that
the average cost in this industry would best be estimated using a quadratic
regression model.
b. The second column of Table 16.3 shows the regression results for the linear
regression model: AC = β 0 + β1Output + ε . The linear regression
model provides a poor fit, which is not surprising given the scatterplot in
Figure 16.2. Not only is Output statistically insignificant, the adjusted R 2
is negative. In order to estimate a quadratic regression model, we have to
first create data on the squared Output variable. A portion of these data,
computed by squaring Output, are shown in Table 16.2. The third column
of Table 16.3 shows the regression results for the quadratic regression
model: A C = β 0 + β 1 Output + β 2Output2 + ε . In the quadratic regression
model, Output is now significant. In addition, the slope coefficient o f Output2
is positive and significant, indicating a U-shaped relationship between average
cost and output.
TABLE 16.3 Estimates o f th e Linear and the Quadratic Regression Models for
Example 16.1
* r e p r e s e n ts s ig n if ic a n c e a t t h e 5 % le v e l. T h e la s t r o w p r e s e n ts a d ju s te d R2 f o r m o d e l c o m p a r is o n .
EXAMPLE 1 6.2
Use the quadratic regression model fro m Example 16.1 to find the output level that
m inim izes costs.
SOLUTION: Given b 1 = − 0.3073 and b2 = 0.0210, the output level that m inimizes
Figure 16.2, this amount closely approximates what we w ould have estimated using
the graph.
23.66 12 49
a. Plot hourly wage against age and evaluate whether a linear or quadratic regres
sion model better captures the relationship. Verify your choice by using the
appropriate numerical measure.
b. Use the appropriate model to predict hourly wages for someone with 16 years
of education and age equal to 30, 50, or 70.
c. According to the model, at what age will someone with 16 years of education
attain the highest wages?
SOLUTION:
a. Figure 16.3 shows a scatterplot of Age against Wage. We superimpose linear and
quadratic trends on the scatterplot. It seems that the quadratic regression model
provides a better fit for the data as compared to the linear regression model.
N OTES: P a r a m e te r e s t im a te s a r e in t h e m a in b o d y o f t h e t a b le w i t h t h e p - v a lu e s in p a r e n th e s e s ; N A d e n o te s n o t
a p p lic a b le ; * r e p r e s e n ts s ig n if ic a n c e a t t h e 5 % le v e l. T h e la s t r o w p r e s e n ts a d ju s t e d R2 f o r m o d e l c o m p a r is o n .
S im ila rly, the predicted h ou rly wage fo r a 50- and a 70-year-old person is
$31.59 and $26.67, respectively. Note that, consistent w ith the estimates, the
h o u rly wage increases as a person ages fro m 30 to 50, but then decreases as a
person ages fro m 50 to 70.
c. In part b, we predicted the h ou rly wage fo r a 30-, 50-, and 70-year-old person
w ith 16 years o f education. Therefore, o f the three ages considered, a 50-year-
old person earns the highest wage. In Figure 16.4, we p lo t the predicted wage
w ith 16 years o f education and vary age fro m 20 to 80 w ith increm ents o f 1. In
order to determ ine the o ptim al age at w hich wage is m axim ized, we also solve
is a b o u t 51 y e a rs , w it h a w a g e o f a b o u t $ 3 1 .6 0 . I t is w o r th n o tin g th a t a t a d if f e r
e n t e d u c a tio n le v e l, p re d ic te d w a g e s w i l l n o t b e th e sa m e , y e t th e h ig h e s t w a g e
w i l l s t ill b e a c h ie v e d a t th e s a m e 51 y e a rs o f a g e . W e a d v is e y o u to p lo t a s im ila r
g ra p h w it h 12 y e a rs o f e d u c a tio n a n d v a r y in g a g e le v e ls .
T h e q u a d r a tic r e g re s s io n m o d e l a llo w s o n e s ig n c h a n g e o f th e s lo p e c a p tu r in g th e in f lu
e n c e o f x o n y . I t is a s p e c ia l c a s e o f a p o ly n o m ia l re g re s s io n m o d e l. P o ly n o m ia l re g re s
s io n m o d e ls d e s c rib e v a rio u s n u m b e rs o f s ig n c h a n g e s . S o m e tim e s a q u a d r a tic re g re s s io n
m o d e l w it h o n e s ig n c h a n g e is r e fe r re d to as a p o ly n o m ia l re g re s s io n m o d e l o f o rd e r 2 . In
fa c t, a lin e a r re g re s s io n m o d e l is a p o ly n o m ia l re g re s s io n m o d e l o f o rd e r 1, w h ic h , w it h a
c o n s ta n t s lo p e c o e ffic ie n t, d o e s n o t a llo w a n y s ig n c h a n g e .
T h e lin e a r a n d th e q u a d ra tic re g re s s io n m o d e ls a re th e m o s t c o m m o n p o ly n o m ia l re g re s
s io n m o d e ls . S o m e tim e s , re s e a rc h e rs u se a p o ly n o m ia l re g re s s io n m o d e l o f o rd e r 3 , a ls o c a lle d
th e c u b ic re g re s s io n m o d e l. T h e c u b ic re g re s s io n m o d e l a llo w s f o r tw o c h a n g e s in s lo p e .
F IG U R E 1 6 .5 R e p re s e n ta tiv e s h a p e o f a c u b ic re g re s s io n m o d e l:
y = β 0 + β1 x + β2 x2 + β3 X3 + ε
We often apply the cubic regression m odel when estim ating the total cost curve o f a firm .
G enerally, total costs o f a firm increase gradually and then rapidly, as predicted by the law
o f dim inishin g returns. In general, the law o f dim inishin g returns states that when increas
ing amounts o f one factor o f production (generally labor) are em ployed in a production
process along w ith a fixed amount o f another factor (generally capital), then after some
p oint, the resulting increases in output o f the product become sm aller and smaller. A com
mon example is adding more workers to a jo b , such as assembling a car in a factory. A t
some point, adding more workers w ill result in inefficien cy as workers get in each other’s
way o r w ait fo r access to a machine. Producing one more u n it o f output w ill eventually
cost increasingly more, due to inputs being used less and less effectively. We can th in k o f
Figure 16.5 as a firm ’s typ ica l total cost curve where x and y represent a firm ’s output and
total cost, respectively.
A c u b ic r e g re s s io n m o d e l c a n e a s ily b e e s tim a te d w it h in th e fr a m e w o r k o f a lin e a r
r e g re s s io n m o d e l w h e re w e u se y as th e re s p o n s e v a ria b le a n d x , x 2, a n d x 3 as th e e x p la n a
t o r y v a ria b le s . I t is e a s y to e s tim a te a c u b ic r e g re s s io n m o d e l a fte r w e h a v e c re a te d d a ta
o n x 2 a n d x 3 A s b e fo r e , w e c a n c o m p a re p o ly n o m ia l m o d e ls o f v a rio u s o rd e rs o n th e b a s is
o f a d ju s te d R 2.
EXAMPLE 1 6 .4
T a b le 1 6 .6 s h o w s a p o r t io n o f d a ta o n th e t o t a l c o s t ( in $ 1 ,0 0 0 s ) a n d o u tp u t f o r
p r o d u c in g a p a r t ic u la r p r o d u c t; th e e n tir e d a ta s e t c a n b e fo u n d o n th e te x t w e b s ite ,
la b e le d T o t a l C o st. W e a ls o in c lu d e a p o r t io n o f th e s q u a re d o u tp u t a n d c u b e d o u tp u t
v a ria b le s ; th e s e v a ria b le s w i l l b e u s e d in th e e s tim a tio n p ro c e s s .
a. U s e a c u b ic re g re s s io n m o d e l to e s tim a te t o ta l c o s t ( T C ) .
b. P r e d ic t th e t o ta l c o s t i f a f ir m p ro d u c e s 11 u n its o f th e p ro d u c t.
TA B L E 1 6 .6 D a ta fo r E x a m p le 16 .4 o n T o ta l C o s t y a n d O u t p u t x , n = 4 0
33.92 4 16 64
SOLUTION:
a. I n o r d e r t o e s tim a te a c u b ic r e g r e s s io n m o d e l, w e f i r s t c r e a te d a ta o n th e
s q u a r e d o u t p u t a n d th e c u b e d o u t p u t v a r ia b le s (s e e T a b le 1 6 .6 ) . W e th e n e s t i
m a te T C = β 0 + β 1 O u t p u t + β 20 u t p u t 2 + β 3 O u t p u t 3 + ε . T a b le 1 6 .7 s h o w s
th e r e le v a n t r e g r e s s io n r e s u lts . A l l v a r ia b le s a re s ig n if ic a n t a t th e 5 % le v e l.
I n a d d it io n , th e a d ju s te d R 2 ( n o t s h o w n ) is 0 .8 5 5 1 , w h ic h is h ig h e r th a n th e
a d ju s te d R 2 o f 0 .6 4 5 2 a n d 0 .8 2 8 9 f o r th e lin e a r a n d th e q u a d r a t ic r e g r e s s io n
m o d e ls , r e s p e c tiv e ly . W e a d v is e y o u t o v e r if y th e s e r e s u lts .
T A B L E 1 6 .7 E s tim a te s o f t h e C u b ic R e g re s s io n M o d e l f o r E x a m p le 16.4
N OTES: p - v a lu e s a r e in p a r e n th e s e s ; * r e p r e s e n ts s ig n if ic a n c e a t t h e 5 % le v e l.
b. W e c a lc u la t e th e t o t a l c o s t o f p r o d u c t io n t o a f i r m th a t p r o d u c e s 11 u n it s o f
th e p r o d u c t a s
EXERCISES 16. 1
Concepts 2. Consider the follow ing three models:
For each o f th e e stim a te d m odels, p re d ic t y w h e n x equals For each o f the estimated models, predict y when x equals
X2 NA − 0.31 − 0.27
β 2Size2 + ε.
b. Justify the inclusion of the quadratic term in the model.
X3 NA NA − 0.03
c. Find the predicted num ber o f bids for a firm that has
a bid premium o f 1.2 and firm size o f $ 4 , 8 , 12, and
R2 0.640 0.697 0.698
16 billion, respectively. What firm size is likely to get the
Adjusted R2. 0.636 0.691 0.689
highest num ber o f bids?
a. P re d ic t y f o r x = 2 and 3 w ith each o f th e estim ated 9. F I L E You collect data on 26 m etropolitan areas to analyze
models. average m onthly debt payments in terms o f income and the
b. Select th e m ost a p p ro p ria te m odel. Explain . unem ploym ent rate. A portion o f the data is shown in the
accompanying table; the com plete data set can be found on a. Estimate the model Debt = β 0 + β 1 lnc + β 2Unemp + ε .
the text website, labeled Debt Payments. Is unem ploym ent significant at the 5% level?
M etropolitan Income b. You are told that the unem ploym ent rate m ig ht have a
Area (in $1,000s) Unem ploym ent Debt quadratic influence on m onthly debt payments. Provide
Washington, D.C. $103.50 6.3% $1,285 an intuitive justification for this claim.
SOURCE: e F a n n ie M a e .c o m : b ls .c o m : a n d E x p e r ia n .c o m .
where e ≈ 2.718 is a constant and x is the fu nctio n argument. We can use Excel o r a cal
culator to compute, fo r example, e2 = 7.39, o r e5 = 148.41.
The inverse o f the exponential fu nctio n is the natural logarithm (o r sim ply, log ), that
is, the logarithm w ith the base e ≈ 2.718. In other words,
w h e re l n ( y ) is th e n a tu r a l lo g o f y . F o r e x a m p le , i f y = e2 = 7 .3 9 , th e n ln ( y ) = In ( 7 .3 9 ) = 2 .
S im ila r ly , i f y = e5 = 1 4 8 .4 1 , th e n ln ( y ) = In ( 1 4 8 .4 1 ) = 5 . S in c e e x p ( ln ( x ) ) = x , th e
e x p o n e n tia l fu n c t io n is s o m e tim e s r e fe r r e d to as th e a n t i- lo g f u n c tio n . F in a lly , th e lo g o f
a n e g a tiv e o r z e ro v a lu e is n o t d e fin e d . T h e r e fo r e , w e c a n lo g - tr a n s fo r m o n ly th o s e v a lu e s
th a t a re p o s itiv e .
A s m e n tio n e d e a rlie r , in m a n y a p p lic a tio n s , lin e a r it y is n o t ju s t if ia b le . F o r in s ta n c e ,
c o n s id e r a n e s tim a te d lin e a r re g re s s io n o f a n n u a l fo o d e x p e n d itu r e y o n a n n u a l in c o m e
x: ŷ = 9000 + 0 .2 0 .x. A n e s tim a te d s lo p e c o e f f ic ie n t v a lu e o f b 1 = 0 .2 0 im p lie s th a t a
$ 1 ,0 0 0 in c re a s e in a n n u a l in c o m e w o u ld le a d to a $ 2 0 0 in c re a s e in a n n u a l fo o d e x p e n
d itu r e , ir re s p e c tiv e o f w h e th e r th e in c o m e in c re a s e is f r o m $ 2 0 ,0 0 0 to $ 2 1 ,0 0 0 o r f r o m
$ 5 2 0 ,0 0 0 to $ 5 2 1 ,0 0 0 . S in c e w e w o u ld e x p e c t th e im p a c t to b e s m a lle r a t h ig h in c o m e
le v e ls , i t m a y b e m o re m e a n in g fu l to a n a ly z e w h a t h a p p e n s to fo o d e x p e n d itu r e as in c o m e
in c re a s e s b y a c e rta in p e rc e n ta g e r a th e r th a n b y s o m e d o lla r a m o u n t.
I t is c o m m o n to lo g - tr a n s fo r m v a ria b le s th a t a re n a tu r a lly e x p re s s e d in p e rc e n ta g e s ,
s u c h as in c o m e s , h o u s e p ric e s , a n d s a le s . O n th e o th e r h a n d , v a ria b le s s u c h as a g e , e x p e r i
e n c e , a n d s c o re s a re g e n e r a lly e x p re s s e d in t h e ir o r ig in a l f o r m . W e r e ly b o th o n e c o n o m ic
in t u it io n as w e ll as s ta tis tic a l m e a s u re s to f in d th e a p p ro p ria te f o r m f o r th e v a ria b le s .
W e f ir s t illu s t r a t e lo g m o d e ls w it h o n ly o n e e x p la n a to r y v a ria b le , w h ic h w e la te r
e x te n d to a m u ltip le re g re s s io n m o d e l.
A Log-Log Model
In a lo g - lo g m o d e l, b o th th e re s p o n s e v a r ia b le a n d th e e x p la n a to r y v a r ia b le a re tr a n s
fo r m e d in t o n a tu r a l lo g s . W e c a n w r it e th is m o d e l as
w h e r e ln ( y ) is th e lo g - t r a n s f o r m e d re s p o n s e v a r ia b le a n d ln ( x ) is th e lo g - t r a n s f o r m e d
e x p la n a to r y v a r ia b le . W it h th e s e tr a n s f o r m a t io n s , th e r e la t io n s h ip b e tw e e n y a n d x is
c a p tu r e d b y a c u r v e w h o s e s h a p e d e p e n d s o n th e s ig n a n d m a g n itu d e o f th e s lo p e c o e f
f ic ie n t β 1 . F ig u r e 1 6 .6 s h o w s a c o u p le o f r e p r e s e n ta tiv e s h a p e s o f a lo g - lo g r e g r e s s io n
m o d e l.
F o r 0 < β1 < 1, the log -lo g m odel im plies a positive relationship between x and E(y ) ; as x
increases, E(y) increases at a slower rate. This may be appropriate in the earlier example,
where we expect food expenditure to react positively to changes in income, but w ith the
im pact dim inishin g at higher incom e levels. I f β1 < 0, it suggests a negative relationship
between x and E (y ) ; as x increases, E (y ) decreases at a slower rate. F ina lly, β1 > 1 im plies
a positive and increasing relationship between x and y; this case is less com m on in a log-
log application and is not shown in Figure 16.6. For any application, the estimated value
o f β1 is determ ined by the data.
Note that w h ile the log -lo g regression m odel is nonlinear in the variables, it is s till
linear in the coefficients, thus satisfying the requirem ent o f the linear regression model.
The only requirem ent is that we have to first transform both variables in to logs before run
ning the regression. We should also p o in t out that in a log -lo g regression m odel, the slope
coefficient β1 measures the percentage change in y fo r a given (sm all) percentage change
in x. In other words, β1 is a measure o f elasticity. For instance, i f y represents the quantity
demanded o f a p articular good and x is its u n it price, then β1 measures the price e lasticity
o f demand, a parameter o f considerable econom ic interest. Suppose β1 = − 1.2; then a
1% increase in the price o f this good is expected to lead to a 1.2% decrease in its quantity
demanded.
F inally, even though the response variable is transform ed into loss, we s till make pre
dictions in regular units. Given you may be tempted to use the anti
log function, to make predictions in regular units as
where b0 and b 1 are the co efficie nt estimates. However, this transform ation is know n to
system atically underestimate the expected value o f y. One relatively sim ple correction is
to make predictions as ŷ = exp(b0 + b 1In(x) + se2/ 2), where se is the standard error o f the
estimate fro m the log -lo g m odel. T his correction is easy to im plem ent since v irtu a lly all
statistical packages report se.
THE LO G -L O G REGRESSION M O D E L
L ik e th e lo g - lo g m o d e l, th is m o d e l im p lie s th a t a n in c re a s e in x w i l l le a d to a n in c re a s e
(β1 , > 0 ) o r d e c re a s e (β 1 < 0 ) in E (y ) a t a d e c re a s in g ra te . T h e s e m o d e ls a re e s p e c ia lly
a ttr a c tiv e w h e n o n ly th e e x p la n a to r y v a r ia b le is b e tte r c a p tu re d in p e rc e n ta g e s . F ig u r e 1 6 .7
h ig h lig h t s s o m e r e p re s e n ta tiv e s h a p e s o f th is m o d e l. S in c e th e lo g - lo g a n d th e lo g a r it h m ic
m o d e l c a n a llo w s im ila r s h a p e s , th e c h o ic e b e tw e e n th e tw o m o d e ls c a n b e t r ic k y . W e w i l l
c o m p a re m o d e ls la te r in th is s e c tio n .
THE L O G A R IT H M IC M O D E L
EXAMPLE 1 6 .6
C o n tin u in g w it h th e e a r lie r e x a m p le o f fo o d e x p e n d itu r e , le t th e e s tim a te d re g re s
s io n b e ŷ = 12 + 5 6 6 In (x ).
a. W h a t is th e p re d ic te d fo o d e x p e n d itu re f o r a n in d iv id u a l w h o s e in c o m e is $ 2 0 ,0 0 0 ?
b. W h a t is th e p re d ic te d v a lu e i f in c o m e in c re a s e s to $ 2 1 ,0 0 0 ?
c. In te r p r e t th e s lo p e c o e ffic ie n t, b 1 = 5 6 6 .
S O L U T IO N : F o r th is lo g a r it h m ic m o d e l, w e m a k e p r e d ic tio n s a s ŷ = b 0 + b1 In (x ).
a. T h e p re d ic te d fo o d e x p e n d itu r e f o r a n in c o m e o f $ 2 0 ,0 0 0 is c a lc u la te d as
ŷ = 12 + 5 6 6 × ln ( 2 0 ,0 0 0 ) = $ 5 ,6 1 7 .
b. F o r x = 2 1 ,0 0 0 , w e fin d ŷ = 12 + 5 6 6 × ln ( 2 1 ,0 0 0 ) = $ 5 ,6 4 5 .
c. W ith a 5 % in c re a s e in in c o m e fr o m $ 2 0 ,0 0 0 to $ 2 1 ,0 0 0 , ŷ in c re a s e s f r o m $ 5 ,6 1 7
to $ 5 ,6 4 5 , o r b y a b o u t $ 2 8 . T h is is c o n s is te n t w it h th e in te r p r e ta tio n o f th e
s lo p e c o e f f ic ie n t ; th a t is , b1 = 5 6 6 im p lie s th a t a 5 % in c re a s e in in c o m e w ill
THE E X P O N E N T IA L M O D E L
EXAMPLE 16.7
C ontinuing again w ith the example o f expenditure on food, le t the estimated
re g re s s io n be w ith the standard error o f the estimate,
se= 0 .20 .
a. W h a t is the p re d icte d e xp e n d itu re on fo o d fo r an in d iv id u a l w hose in co m e is
$ 2 0 ,0 0 0 ?
b. W h a t is the p re d icte d va lu e i f in co m e increases to $21,000?
c. In te rp re t the slope c o e ffic ie n t, b 1 = 0.00005.
a. P lo t r e n t a g a in s t e a c h o f th e th re e e x p la n a to r y v a ria b le s a n d e v a lu a te w h e th e r
th e r e la tio n s h ip is b e s t c a p tu re d b y a lin e o r a c u rv e . I d e n t if y v a ria b le s th a t
m a y r e q u ir e a lo g - tr a n s fo r m a tio n .
b. E s tim a te th e lin e a r a n d th e r e le v a n t lo g m o d e ls to p r e d ic t r e n t f o r a
1,6 0 0 - s q u a re - fo o t h o m e w it h th re e b e d ro o m s a n d tw o b a th ro o m s .
SOLUTION: Given the nature o f Beds and Baths, we w ill specify these variables
only in regular units. We w ill, however, consider l og-transform ations fo r Rent and
Sqft, since th eir changes are often expressed in percentages.
a. In Figure 16.9, we p lo t Rent against (a) Beds and (b) Baths and superimpose
linear and exponential curves (recall that an exponential m odel log-transform s
o nly the response variable).
FIGURE 16.9 Com paring Rent against (a) Beds and (b) Baths
It is hard to te ll fro m Figure 16.9 whether the relationship between Rent and
Beds or Rent and Baths is better captured by a line or a curve. We w ill use
fo rm al num erical measures fo r the selection.
We now p lo t Rent against Sqft in Figure 16.10.
H e re i t a p p e a rs th a t th e r e la tio n s h ip b e tw e e n R e n t a n d S q ft is b e tte r c a p tu re d
b y a c u rv e th a n a lin e . F ig u r e 1 6 .1 0 s h o w s th a t a lo g a r it h m ic m o d e l th a t lo g -
tr a n s fo r m s S q ft fit s th e d a ta b e tte r th a n th e lin e a r m o d e l, s u g g e s tin g th a t as
square footage increases, rent increases at a decreasing rate. In other words,
the increase in Rent is higher when Sqft increases from 1,000 to 2,000 than
from 2,000 to 3,000. Two other models w orth considering are the exponential
m odel, where only Rent is log-transform ed, and a log -lo g m odel, where both
Rent and Sqft are log-transform ed. In order to avoid a “ cluttered” figure, these
curves are not superimposed on the scatterplot; however, we w ill fo rm a lly
evaluate a ll models.
b. W h ile the above visual tools are instructive, we evaluate fo u r models and use
num erical measures to select the m ost appropriate m odel fo r prediction.
In order to estimate these models, we firs t log-transform Rent and S qft; see the
last tw o colum ns o f Table 16.9.
T A B L E 1 6 .9 T r a n s fo rm in g R e n t a n d S q ft in t o Lo g s
I n M o d e ls 1 a n d 2 , w e u s e R e n t as th e re s p o n s e v a r ia b le w it h B e d s a n d B a th s ,
a lo n g w it h S q ft in M o d e l 1 a n d ln ( S q f t ) in M o d e l 2 , as th e e x p la n a to r y v a r i
a b le s . S im ila r ly , in M o d e ls 3 a n d 4 , w e u s e ln ( R e n t) as th e re s p o n s e v a ria b le
w it h B e d s a n d B a th s , a lo n g w it h S q ft in M o d e l 3 a n d I n ( S q f t) in M o d e l 4 , as
th e e x p la n a to r y v a ria b le s . M o d e l e s tim a te s a re s u m m a r iz e d in T a b le 1 6 .1 0 .
T A B L E 1 6 .1 0 R e g re s s io n R e s u lts fo r E x a m p le 16.8
a t th e 5 % le vel.
F o r th e m o s t p a rt, th e n u m b e r o f b e d ro o m s a n d th e s q u a re fo o ta g e o f th e h o u s e
a re s t a t is t ic a lly s ig n ific a n t a t th e 5 % le v e l, w h ile th e n u m b e r o f b a th r o o m s is
in s ig n if ic a n t . W e u s e th e m o d e l r e s u lts to p r e d ic t r e n t f o r a 1 ,6 0 0 - s q u a r e - fo o t
h o m e w it h th re e b e d ro o m s a n d tw o b a th ro o m s . I n o r d e r to m a k e a p r e d ic tio n
w it h M o d e ls 3 a n d 4 , w h ic h a re b o th b a s e d o n ln ( R e n t) , w e m u s t a d d th e
c o r r e c tio n te rm se2/ 2 .
C o m p a r in g L in e a r a n d L o g -T r a n s fo r m e d M o d e ls LO 1 6 .3
As seen in Example 16.8, it is often not clear w hich regression model is best suited fo r an D e s c rib e t h e m e t h o d
application. W h ile we can use economic intuition and scatterplots fo r direction, we also used to c o m p a re
ju s tify our selection on the basis o f numerical measures. In Chapter 14, we introduced R2 lin e a r w i t h lo g
to compare models based on the same number o f explanatory variables; we c o m p a r e d a d t r a n s f o r m e d m o d e ls .
justed R2 i f the number o f explanatory variables was different. Such comparisons are v a l i d
o nly when the response variable o f the competing models is the same. Since R2 m e a s u r e s
the percentage o f sample variations o f the response variable explained by the model, we
cannot compare the percentage o f explained variations o f y w ith that o f ln(y). Comparing
models based on the computer-generated R2 is like comparing apples w ith oranges. For
a valid comparison, we need to compute the percentage o f explained variations o f y even
though the estimated model uses ln(y) as the response variable. To do this, it w ill help to
revisit the form ula fo r calculating R2. Recall that an easy way to compute R2 is by squar
ing the sample correlation coefficient o f y and ŷ .
R E V I S I T IN G THE C A L C U L A T I O N OF THE
C O E F F I C I E N T O F D E T E R M I N A T I O N R2
EXAMPLE 16.9
Revisit the fo ur regression models in Example 16.8 and determine w hich model is
best suited fo r making predictions.
E xcel is useful in perform ing these calculations— it provides the predicted val
ues fo r the ln(R ent) i f we check Residuals in the Regression dialog box. Since
Excel provides fo r M odel 4, we can easily compute the predicted rent as
I n T a b le 16.11, we present a portion o f these calcula
y
645 6.4778 655.7334
675 6.6007 742.5210
We use the C orrelation fu nctio n in Excel (select Formulas > Insert Function >
C O R R E L) to derive the correlation between y and ŷ , in colum ns 1 and 3 o f
Table 16.11, as =
ŷry 0.8691. We square this value to compute the co efficie nt o f
determ ination, R2 = (0.8691)2 = 0.7553. We can now compare this value w ith the
computer-generated value fo r M odel 2. We conclude that M odel 2 is better suited
fo r m aking predictions, since 0.8482 > 0.7553.
S Y N O P S IS OF IN T R O D U C T O R Y CASE
T h e r e c e s s io n - r e s is ta n c e o f c a m p u s t o w n s h a s p r o m p t e d m a n y a n a ly s ts
t o c a ll in v e s t m e n t in o f f - c a m p u s s t u d e n t h o u s in g a s m a r t c h o ic e (T he
W a ll S tre e t J o u r n a l, S e p t e m b e r 2 4 , 2 0 1 0 ). F irst, t h e r e is a s ta b le s o u rc e o f
d e m a n d in c o lle g e t o w n s , as c a s h - s t r a p p e d p u b lic u n iv e r s it ie s a re u n
a b le t o h o u s e a ll s t u d e n t s . S e c o n d , t h is d e m a n d m a y a c t u a lly im p r o v e
d u e t o a p r o je c t e d in c r e a s e in c o lle g e e n r o llm e n t . In t h is s tu d y , A n n
A r b o r , w h ic h is h o m e t o t h e m a in c a m p u s o f t h e U n iv e r s ity o f M ic h i
g a n , is u s e d t o s t u d y r e n t a l o p p o r t u n it ie s . F o u r r e g r e s s io n m o d e ls a n a
ly z e t h e m o n t h l y r e n t (R e n t) o n t h e b a s is o f t h e n u m b e r o f b e d r o o m s
(B e d s), t h e n u m b e r o f b a t h r o o m s (B a th s ), a n d t h e s q u a r e f o o t a g e (S q ft)
o f o f f - c a m p u s h o u s e s . N o n lin e a r it ie s b e t w e e n t h e v a r ia b le s a re c a p t u r e d b y t r a n s f o r m
in g R e n t a n d / o r S q ft in t o n a t u r a l lo g s . T h e c o e f f ic ie n t o f d e t e r m in a t io n R2, c o m p u t e d in
t h e o r i g in a l u n its , is u s e d t o s e le c t t h e b e s t m o d e l. T h e s e le c te d m o d e l is e s t im a t e d as
T h e b e d r o o m c o e f f i
c ie n t im p lie s t h a t f o r e v e r y a d d it io n a l b e d r o o m , t h e m o n t h l y r e n t is p r e d ic t e d t o g o u p
b y a b o u t $ 1 3 2 , h o ld in g o t h e r f a c t o r s c o n s t a n t . S im ila r ly , f o r e v e r y o n e p e r c e n t in c r e a s e in
s q u a r e f o o t a g e , t h e m o n t h l y r e n t is p r e d ic t e d t o in c re a s e b y a b o u t $ 6 .7 5 ( 6 7 5 / 1 0 0 ) . T h is
s a m p le re g r e s s io n m o d e l c a n a ls o b e u s e d t o m a k e p r e d ic t io n s f o r r e n t . F o r e x a m p le , a
1 , 0 0 0 - s q u a r e - fo o t h o u s e w i t h t w o b e d r o o m s a n d o n e b a t h r o o m is p r e d ic t e d t o r e n t f o r
$ 1 ,0 5 5 . S im ila r ly , a 1 , 6 0 0 - s q u a r e - fo o t h o u s e w i t h t h r e e b e d r o o m s a n d t w o b a t h r o o m s
is p r e d ic t e d t o r e n t f o r $ 1 ,5 4 0 . T h e s e re s u lts a re u s e fu l t o a n y in v e s t o r in t e r e s t e d in o f f -
c a m p u s h o u s in g in A n n A r b o r .
EXERCISES 16.2
C oncepts 1 5. C o n s id e r t h e f o l lo w in g 1 0 o b s e r v a t io n s o f y a n d x .
34.62 8.06 12.67 23.02 11.82 27.23 18.23 11.23 11.00 21.07
10. Consider the follow ing four estimated models:
22
2 11 19 2 21 18 11 19 20
a. P lo t t h e a b o v e d a t a t o c h o o s e b e t w e e n t h e lin e a r a n d
t h e e x p o n e n t ia l m o d e l.
b. J u s t if y y o u r c h o ic e u s in g R2 d e f in e d in te r m s o f y .
c. W it h t h e b e s t - f i t t i n g m o d e l, p r e d ic t y f o r x = 2 0 .
a. I n t e r p r e t t h e s lo p e c o e f f i c ie n t in e a c h o f th e s e e s tim a t e d
m o d e ls .
A p p lic a tio n s
b. F o r e a c h m o d e l, w h a t is t h e p r e d ic t e d c h a n g e in y w h e n
1 6. F I L E A c c o r d in g t o t h e W o r l d H e a l t h O r g a n i z a t i o n , o b e s it y
x in c re a s e s b y 1 % , f r o m 1 0 0 t o 1 01 ?
h a s r e a c h e d e p id e m ic p r o p o r t i o n s g lo b a lly . W h ile o b e s it y
1 1. C o n s id e r t h e f o l lo w in g e s t im a t e d m o d e ls : h a s g e n e r a lly b e e n lin k e d w i t h c h r o n ic d is e a s e a n d d is a b ility ,
re s e a r c h e rs a r g u e t h a t i t m a y a ls o a f f e c t w a g e s . B o d y M a s s
I n d e x ( B M I) is a w id e ly u s e d w e i g h t m e a s u r e t h a t a ls o a d ju s ts
f o r h e ig h t . A p e r s o n is c o n s id e r e d n o r m a l w e i g h t i f B M I is
b e t w e e n 1 8 .5 t o 2 5 , o v e r w e ig h t i f B M I is b e t w e e n 2 5 t o 3 0 ,
a n d o b e s e i f B M I is o v e r 3 0 . T h e a c c o m p a n y in g t a b le s h o w s
a. I n t e r p r e t t h e s lo p e c o e f f i c ie n t in e a c h o f t h e s e e s t im a t e d a p o r t io n o f d a t a o n t h e s a la r y o f 3 0 c o lle g e - e d u c a t e d m e n
m o d e ls . w i t h t h e i r r e s p e c tiv e B M I; t h e c o m p l e t e d a ta , la b e le d B M I ,
b. F o r e a c h m o d e l, w h a t is t h e p r e d ic t e d c h a n g e in y w h e n a re a v a ila b le o n t h e t e x t w e b s ite .
b. E s tim a te a n e x p o n e n t ia l m o d e l u s in g lo g o f s a la r y as t h e
W h a t is t h e e s t im a t e d s a la r y o f a c o lle g e - e d u c a t e d m a n
13. Consider th e sam ple regressions fo r th e linear, th e
w i t h a B M I o f 2 5 ? W it h a B M I o f 30?
lo g a rith m ic, th e e xp o ne n tia l, and th e lo g -lo g models. For
c. W h ic h o f t h e a b o v e t w o m o d e ls is m o r e a p p r o p r ia t e f o r
each o f th e e stim ated m odels, p re d ic t y w h e n x equals 50 .
t h is a p p lic a t io n ? U se R2 f o r c o m p a r is o n .
R e s p o n s e V a r ia b le : y R e s p o n s e V a r ia b le : ln (y )
1 7. F I L E P ro fe s s o r O r le y A s h e n f e lt e r o f P r in c e to n U n iv e r s it y
M odel 1 M odel 2 M odel 3 M odel 4
is a p io n e e r in t h e f i e ld o f w in e e c o n o m ic s . H e c la im s
In te r c e p t 1 8 .5 2 − 6 .7 4 1 .4 8 1 .0 2
t h a t , c o n t r a r y t o o ld o r t h o d o x y , t h e q u a l it y o f w in e c a n
x 1 .6 8 NA 0 .0 6 NA b e e x p la in e d m o s t ly in te r m s o f w e a t h e r c o n d it io n s . W in e
ln ( x ) NA 2 9 .9 6 NA 0 .9 6 r o m a n tic s a c c u s e h im o f u n d e r m in i n g t h e w h o l e w in e -
t a s t in g c u lt u r e . In a n in t e r e s t in g c o - a u t h o r e d p a p e r t h a t
2 3 .9 2 19.71 0 .1 2 0 .1 0 a p p e a r e d in C h a n c e m a g a z in e in 1 9 9 5 , h e ra n a m u lt ip l e
Se
r e g r e s s io n m o d e l w h e r e q u a lit y , m e a s u r e d b y t h e p ric e s
14. Consider th e fo llo w in g 10 o b servations o f y and x . t h a t w in e s f e t c h a t a u c t io n s , is u s e d as t h e r e s p o n s e v a r ia b le
22.21 21.94 22.83 22.66 21.44 22.51 22.87 22.50 22.88 23.16 y . T h e e x p l a n a t o r y v a r ia b le s u s e d in t h e a n a ly s is w e r e t h e
12 5 15 16 4 8 11 12 18 16 a v e r a g e t e m p e r a t u r e in C e ls iu s x 1, t h e a m o u n t o f w i n t e r ra in
x 2/ t h e a m o u n t o f h a r v e s t r a in x 3, a n d t h e y e a rs s in c e v in t a g e
a. P lo t t h e a b o v e d a ta t o c h o o s e b e t w e e n t h e lin e a r a n d
x 4. A p o r t io n o f t h e d a t a is s h o w n in t h e a c c o m p a n y in g ta b le ;
t h e l o g a r it h m i c m o d e ls .
t h e c o m p l e t e d a ta , la b e le d W i n e P r ic in g , is a v a ila b le o n t h e
b. J u s tify y o u r c h o ic e u s in g t h e a p p r o p r ia te n u m e r ic a l m e a s u re .
t e x t w e b s ite .
c. U se t h e s e le c te d m o d e l t o p r e d i c t y f o r x = 10.
Winter Harvest Male Life Female Life People/
Price Temperature Rain Rain Vintage Country Expectancy Expectancy Physician
1.4448 17.12 600 160 31 Argentina 67 74 370
1.8870 16.73 690 80 30 Bangladesh 54 53 6,166
a. U se t h e a b o v e d a t a t o e s t im a t e a lin e a r m o d e l, y = β 0 + a. C o n s t r u c t a s c a t t e r p lo t o f f e m a le lif e e x p e c t a n c y
β 1 x 1 + β 2 x 2 + β3 x 3 + β4 x 4 + e . W h a t is t h e p r e d ic t e d a g a in s t t h e p e o p le - t o - p h y s ic ia n r a tio . S u p e r im p o s e a
p r ic e i f x 1 = 16, x 2 = 6 0 0 ,x 3 = 120, a n d x 4 = 20? lin e a r t r e n d a n d a l o g a r it h m i c t r e n d t o d e t e r m i n e t h e
b. U se t h e a b o v e d a ta t o e s tim a t e a n e x p o n e n t ia l m o d e l, a p p r o p r ia t e m o d e l.
ln ( y ) = β0 + β 1 x 1 + β 2 x 2 + β3 x 3 + β4 x 4 + ε . W h a t is t h e b. E s tim a te a s im p le lin e a r r e g r e s s io n m o d e l w i t h lif e
p r e d ic te d p r ic e i f x 1 = 1 6 , x 2 = 6 0 0 , x 3 = 1 2 0 , a n d x 4 = 2 0 ? e x p e c t a n c y o f f e m a le s a s t h e r e s p o n s e v a r ia b le a n d t h e
c. U se R2 t o s e le c t t h e a p p r o p r ia t e m o d e l f o r p r e d ic t io n . p e o p l e - t o - p h y s ic i a n r a t i o as t h e e x p l a n a t o r y v a r ia b le .
t h a t a p e r s o n is e x p e c t e d t o liv e . T h e r e is a h u g e v a r ia t io n
A n i m p o r t a n t f a c t o r f o r s u c h v a r ia b il i t y is t h e a v a i l a b i li t y o f in d ic e s .
s u it a b le h e a lt h c a re . O n e m e a s u r e o f a p e r s o n 's a c c e s s t o
E s tim a te ln ( Q ) = β 0 + β 1 In(L ) + β 2 In (K ) + ε .
h e a lt h c a re is t h e p e o p le - t o - p h y s ic ia n r a tio . W e e x p e c t life
e x p e c t a n c y t o b e lo w e r f o r c o u n t r ie s w h e r e t h is r a t io is h ig h . a. W h a t is t h e p r e d ic t e d c h a n g e in o u t p u t i f la b o r in c re a s e s
T h e a c c o m p a n y in g t a b le lis ts a p o r t io n o f lif e e x p e c t a n c y o f b y 1 % , h o l d in g c a p it a l c o n s t a n t ?
m a le s a n d fe m a le s in 4 0 c o u n t r ie s a n d t h e i r c o r r e s p o n d in g b. H o ld in g c a p ita l c o n s t a n t , c a n w e c o n c lu d e a t t h e 5 %
p e o p le - t o - p h y s ic ia n r a t io ; t h e c o m p l e t e d a t a , la b e le d L i f e le v e l t h a t a 1 % in c r e a s e in la b o r w i l l in c r e a s e t h e o u t p u t
E x p e c t a n c y , a re a v a ila b le o n t h e t e x t w e b s ite . b y m o r e t h a n 0 .5 % ?
WRITING WITH STATISTICS
100 79 31 $105,000
In a survey of 100 working adults, respondents were asked to report their age and family Sample
income, as well as rate their happiness on a scale of 0 to 100. This report summarizes the
analysis of several regression models that examine the influence of age and income on the Report—
perceived happiness of respondents. The models used various transformations to capture
interesting nonlinearities suggested by recent research reports. For example, one such report Understanding
shows that people get happier as they get older, despite the fact that old age is associated with
a loss of hearing, vision, and muscle tone (The New York Times, May 31, 2010). In addition, Happiness
while people start out feeling pretty good about themselves in their 20s, their self-assessed
happiness deteriorates until around age 50 and then improves steadily thereafter. In order
to quantify this possible quadratic effect, both age and age-squared variables are used for
the regression. Also, the natural log of income is considered in order to capture the possible
diminishing effect on happiness of incomes above $75,000 (Time Magazine, September 6,
2010). The results of the various regression models are summarized in Table 16. A.
T A B L E 1 6 . A R e g r e s s io n R e su lts
NOTES: Param eter estim ates are in th e to p portion o f th e tab le w ith th e p-values in parentheses; NA d e n o te s n o t applicable;
*represents significance a t th e 5% level.The last row presen ts th e adjusted R!2values for m odel com parison.
Model 4 was selected as the most appropriate model because it has the highest adjusted R 2
value of 0.6907. The estimated parameters of this model were used to make predictions.
For instance, with family income equal to $80,000, the predicted happiness for a 30-, 50-,
and 70-year-old is 79.09, 69.00, and 78.17, respectively. Note that these results are con
sistent with those suggesting that happiness first decreases and then increases with age.
Specifically, using the estimated coefficients for Age, a person is least happy at 50.4 years
of age. These results are shown graphically in Figure 16.A(a) where Happiness is plotted
against Age, holding Income fixed at $80,000.
The regression results were also used to analyze the income effect. For instance, for a
60-year-old, the predicted happiness with family income of $50,000, $75,000, and $100,000
is 65.20, 70.36, and 74.02, respectively. Note that there is a greater increase in Happiness
when income increases from $50,000 to $75,000 than when it increases from $75,000 to
$100,000. These results are shown in Figure 16.A(b) where happiness is plotted against
Income, holding Age fixed at 60 years. Overall, the results support recent research findings.
C oncep tu al Review
LO 1 6 .1 Use and ev a lu a te polynom ial regression m od els.
In a quadratic regression m odel, y = β0 + β1x + β 2x 2 + ε , the sign of the coefficient
β 2 determines whether the relationship between x and E (y ) is U-shaped (β 2 > 0) or
inverted U-shaped (β 2 < 0). P redictions are made by ŷ = b0 + b 1x + b2x 2.
In a quadratic regression model, the marginal effect of x on ŷ is approximated by
b 1 + 2b2x ; so this effect depends on the value of x. It is common to use when
interpreting the m arginal effect. The quadratic equation reaches a m axim um ( if b2 < 0)
or m inim um ( if b2 > 0) at
A c u b ic re g re s s io n m o d e l, y = β 0 + β 1 x + β 2 x 2 + β 3 x 3 + ε , a llo w s tw o s ig n c h a n g e s o f
th e s lo p e c a p tu r in g th e in flu e n c e o f x o n E ( y ) . P r e d ic t io n s a re m a d e b y ŷ = b 0 + b 1x +
b 2x 2 + b 3x 3.
W e c o m p a re p o ly n o m ia l re g re s s io n m o d e ls o f v a rio u s o rd e rs o n th e b a s is o f a d ju s te d R 2.
LO 1 6 .2 U s e a n d e v a lu a t e lo g t r a n s f o r m e d m o d e ls .
In a lo g -lo g m o d e l, ln ( y ) = β 0 + β 1 ln ( x ) + ε , a n d th e s lo p e c o e ffic ie n t β 1 m e a s u re s th e
a p p ro x im a te p e rc e n ta g e c h a n g e in E (y ) w h e n x in c re a s e s b y o n e p e rc e n t. P r e d ic tio n s
a re m a d e b y ŷ = e x p ( b 0 + b1
ln ( x ) + se2/ 2 ) , w h e re b 0 a n d b 1 a re th e c o e f f ic ie n t e s tim a te s
a n d se is th e s ta n d a rd e r r o r o f th e e s tim a te .
In a lo g a r it h m ic m o d e l, y = β 0 + β 1 ln ( x ) + ε , a n d β 1 / 100 m e a s u re s th e a p p ro x im a te
c h a n g e in E (y ) w h e n x in c re a s e s b y o n e p e rc e n t. P r e d ic tio n s a re m a d e b y ŷ = b0 +
b 1ln ( x ) w h e re b 0 a n d b 1 a re th e c o e f f ic ie n t e s tim a te s .
In a n e x p o n e n tia l m o d e l, ln ( y ) = β0 + β 1 x + ε , and β1 × 1 0 0 m e a s u re s th e a p p r o x i
m a te p e rc e n ta g e c h a n g e in E (y ) w h e n x in c re a s e s b y o n e u n it. P r e d ic t io n s a re m a d e b y
ŷ = e x p ( b 0 + b1x + se2/ 2 ) , w h e re b 0 a n d b 1 a re th e c o e f f ic ie n t e s tim a te s a n d se is th e
s ta n d a rd e r r o r o f th e e s tim a te .
LO 1 6 .3 D e s c r ib e t h e m e t h o d u s e d t o c o m p a r e lin e a r w it h lo g t r a n s f o r m e d m o d e ls .
W e u se th e c o e ffic ie n t o f d e te r m in a tio n R 2 to c o m p a re m o d e ls th a t e m p lo y th e sa m e n u m
b e r o f e x p la n a to ry v a ria b le s a n d u se a d ju s te d R 2 i f th e n u m b e r d iffe r s . S u c h c o m p a ris o n s
a re v a lid o n ly w h e n th e re s p o n s e v a ria b le o f th e c o m p e tin g m o d e ls is th e s a m e . In o th e r
w o rd s , w e c a n n o t c o m p a re th e p e rc e n ta g e e x p la in e d v a ria tio n s o f y w it h th a t o f ln ( y ). F o r
a v a lid c o m p a ris o n , f o r a n y m o d e l th a t uses l n ( y ) as th e re s p o n s e v a ria b le , w e c o m p u te R 2
as R 2 = ( r yŷ) 2, w h e re r yŷ is th e s a m p le c o rr e la tio n c o e ffic ie n t b e tw e e n y a n d ŷ .
a. Estimate and in te rp re t th e m odel: Salary = β 0 + β1PC + a. Estimate the linear model, Price = β 0 + β 1 Sqft +
β 2TD + β3A
g e + e. Show th a t th is m odel is preferable to β 2Beds + β 3Baths + ε , and the exponential model,
a m odel th a t uses log o f salary as th e response variable. In (Price) = β 0 + β 1 Sqft + β 2Beds + β 3Baths + ε .
b. Consider th e qu a dra tic e ffe ct o f Age by a d d in g Age2 in b. Interpret the slope coefficients o f the estimated models.
th e regression. Use a p a rtia l F t e s t to d e te rm in e th e jo in t c. Use the coefficient o f determ ination to choose the
statistical significance o f Age and A ge2. preferred model.
24. F I L E A n u tritio n is t w ants to understand th e influence o f 25. F I L E savings rate has declin e d d ra m a tica lly over th e past
incom e and healthy fo o d on th e incidence o f sm oking. He fe w decades (CNNMoney.com, June 3 0 , 2010). W h ile some
collects 2009 data on th e percentage o f smokers in each state econom ists are e x tre m e ly concerned a b o u t th is decline,
in th e U.S. and th e corresponding m edian incom e and the o thers believe th a t it is a nonissue. Consider th e fo llo w in g
percentage o f th e po p ula tio n th a t regularly eats fruits and m o n th ly data on th e personal savings rate (Savings) and
vegetables. A p o rtio n o f th e data is shown below ; th e com plete th e personal disposable in co m e (Incom e) in th e U.S. fro m
data set can be fo u nd on th e te x t website, labeled Smoking. January 2007 to N o ve m b e r 2010; th e co m p le te dataset,
CASE STUDIES
C a s e S tu d y 16.1
E x e c u tiv e c o m p e n s a tio n h a s ris e n d r a m a tic a lly b e y o n d th e r is in g le v e ls o f a n a v e ra g e
w o r k e r ’ s w a g e o v e r th e y e a rs . T h is h a s b e e n a h o t to p ic f o r d is c u s s io n , e s p e c ia lly w it h
th e c r is is in th e fin a n c ia l s e c to r a n d th e c o n tro v e r s y o v e r th e fe d e ra l b a ilo u t. T h e g o v e rn
m e n t is e v e n c o n s id e r in g a c a p o n h ig h - f ly in g s a la rie s f o r e x e c u tiv e s ( T h e N e w Y o rk T im e s ,
F e b ru a ry 9 , 2 0 0 9 ). C o n s id e r th e f o llo w in g d a ta th a t lin k to ta l c o m p e n s a tio n o f th e 4 5 5
h ig h e s t- p a id C E O s in 2 0 0 6 w it h tw o p e rfo rm a n c e m e a s u re s ( in d u s tr y -a d ju s te d re tu r n o n
a s se ts , R O A , a n d in d u s try - a d ju s te d s to c k r e tu r n ) a n d th e f ir m ’ s s iz e ( T o ta l A s s e ts ). T h e
c o m p le te d a ta , la b e le d E x e c u t iv e C o m p e n s a t i o n , a re o n th e te x t w e b s ite .
S ource : SEC w e b s it e a n d C o m p u s ta t.
I n a r e p o r t, u s e th e s a m p le in f o r m a t io n to :
1. E s tim a te tw o m o d e ls w h e re e a c h m o d e l u se s C o m p e n s a tio n as th e re s p o n s e v a r ia b le
a n d A d j R O A a n d A d j R e tu rn as th e e x p la n a to r y v a ria b le s a lo n g w it h T o ta l A s s e ts in
M o d e l 1 a n d n a tu r a l lo g o f T o ta l A s s e ts in M o d e l 2 .
2. U s e th e p re fe r re d m o d e l to p r e d ic t c o m p e n s a tio n g iv e n th e a v e ra g e v a lu e s o f th e
e x p la n a to r y v a ria b le s .
C a s e S tu d y 1 6 .2
A B r it is h s u rv e y ju s t re v e a le d th a t th e N e w Y o r k Y a n k e e s b a s e b a ll te a m p a y s t h e ir p la y
e rs , o n a v e ra g e , m o re th a n a n y o th e r te a m in th e w o r ld ( h ttp ://s p o r ts illu s tr a te d .c n n .c o m ,
A p r il 7 , 2 0 1 0 ). B re n d a n C o n n o lly , a s ta tis tic ia n f o r a M a jo r L e a g u e B a s e b a ll ( M L B ) te a m .
wants to elaborate on the salary of baseball players. Excluding pitchers from his analysis,
he believes that a baseball player’s batting average (BA), runs batted in (RBI), and years
of experience playing professional baseball (Experience) are the most important factors
that influence a player’s salary. Further, he believes that salaries rise with experience only
up to a point, beyond which they begin to fall; in other words, experience has a quadratic
effect on salaries. Brendan collects data on salary (in $ 1,000s), BA, RBI, and experience
for 138 outfielders in 2008. A portion o f the data is shown in the accompanying table; all
data can be found on the text website, labeled M L B Salary.
Data for Case Study 16.2 M ajor League Baseball O utfielder Data, n = 138
FILE Salary
Player (in $1,000s) BA RBI Experience
1. Nick Markakis 455 299 87 3
2. Adam Jones 390 261 23 3
NOTES: All data collected from usatoday.com or espn.com; BA and RBI are averages over the player's professional life
through 2008. For exposition, BA has been multiplied by 1000.
SOURCE: C e n s u s B u re a u a n d B o a r d o f G o v e r n o r s .
Up until now, regression analysis has allowed us to answer questions such as: W hat is the contribu
tion o f advertisement expenditures to the firm's sales? Can we improve elementary education by
reducing class size? How much will an additional hour o f review before the final exam contribute
to the score? All o f these questions use response and explanatory variables that are quantitative
in nature. There are other im portant applications tha t use qualitative variables representing tw o or
more categories. For instance, we may w a nt answers to questions such as: Do w om en get paid as
much as men for the same work? Are sales o f electronic goods higher in the 4th quarter than in the
other quarters? What is the influence o f family incom e on the probability o f buying a house? In order
to answer these questions, the regression analysis must incorporate qualitative response and/or ex
planatory variables. This chapter examines these kinds o f situations, using methods called d u m m y
variable models and binary choice probability models.
I N T R O D U C T O R Y C A S E
F IL E Salary Experience
Individual (in $ 1,000s) (in years) Gender Age
1 67.50 14 Male Under
2 53.51 6 Male Under
M a r y w o u ld lik e t o u s e t h e s a m p le in f o r m a t io n in T a b le 17.1 t o :
1. T e s t w h e t h e r s a la ry d if f e r s b y a f ix e d a m o u n t b e t w e e n m a le s a n d fe m a le s .
2. D e t e r m in e w h e t h e r t h e r e is e v id e n c e o f a g e d is c r im in a t io n in s a la rie s .
3. D e t e r m in e i f t h e s a la ry d if f e r e n c e b e t w e e n m a le s a n d f e m a le s in c re a s e s w i t h e x p e r ie n c e .
A s y n o p s is o f t h is c a s e is p r o v id e d a t t h e e n d o f S e c tio n 17.2.
17.1 Dummy Variables
In the previous chapters, a ll the variables used in the regression applications have been
quantitative; in other w ords, they assume some num erical values. For instance, the
variables used e arlie r in the regression describing m on thly rents based on the square
footage and the num ber o f bedroom s and bathroom s are a ll quantitative. In e m p irical
w o rk, however, it is very com m on to have some variables that are qualitative in nature.
A lth o u g h qualitative variables can be described by several categories, they are often
binary, m eaning that they represent one o f o n ly tw o possible categories. Exam ples in
clude gender (m ale o r fem ale), home ow nership (own or do not ow n), Internet connec
tio n (yes o r no), and college choice (p u b lic o r private).
Q U A N T I T A T I V E V E R S U S Q U A L I T A T I V E V A R IA B L E S IN R E G R E S S IO N
A D U M M Y VARIABLE
LO 1 7 .1 For the sake o f sim p licity, we w ill begin this section w ith a m odel containing one quan
U s e d u m m y v a r ia b le s
tita tive explanatory variable and one dum m y variable. As we w ill see shortly, the m odel
t o c a p t u r e s s h ift
can easily be extended to include additional variables.
o f t h e in te rc e p t.
C onsider the fo llo w in g m odel:
ŷ = b0 + b 1x.
The dum m y variable d used as an explanatory variable allow s a shift in the intercept o f
the estim ated regression line. N ote that the estim ated intercept b0 w hen d = 0 shifts to
( b0 + b2) w hen d = 1. In other w ords, the dum m y variable enables us to use a single re
gression equation to represent both categories o f the qualitative variable. A ll we need to
do is to use d along w ith the other explanatory variable x in the regression. Figure 17.1
show s a shift in the intercept o f the estim ated regression line w hen d = 0 changes to
d = 1, given b2 > 0.
N ote that the dum m y variable d affects the intercept, but not the slope, o f the linear
regression line. Som etim es, d is referred to as the intercept dummy. Shortly, we will
consider an interaction variable, xd, that affects the slope o f the regression line.
EXAMPLE 17.1
The objective outlined in the introductory case is to determ ine if there is any gender
or age discrim ination at a large liberal arts college. U se the data in Table 17.1 to
answ er the follow ing questions.
a. E stim ate y = β 0 + β1x + β 2d 1 + β3 d2 + ε , w here y is the annual salary (in
$1,000s) o f a professor, x is the num ber o f years o f experience, d 1 is a dum m y
variable that equals 1 if the professor is m ale and 0 otherw ise, and d2 is a
dum m y variable that equals 1 if the professor is 60 years o f age or older and 0
otherw ise.
b. C om pute the predicted salary o f a 50-year-old m ale professor w ith 10 years
o f experience. C om pute the predicted salary o f a 50-year-old fem ale professor
w ith 10 years o f experience. D iscuss the im pact o f gender on predicted salary.
c. C om pute the predicted salary o f a 65-year-old fem ale professor w ith 10 years
o f experience. D iscuss the im pact o f age on predicted salary.
SOLUTION:
a. To estim ate the above m odel, we first convert the qualitative variables in
Table 17.1 to their respective gender and age dum m y variables, d 1and d2,
in Table 17.2.
LO 17.2 Dummy variables are treated just like other explanatory variables; that is, all statistical
Test for d if fe r e n c e s tests discussed in Chapter 15 remain valid. In particular, using the t test we can examine
b e tw e e n th e whether a particular dummy is statistically significant. This allows us to determine if the
ca teg o ries o f a response variable depends on the two categories represented by this dummy. An F test
q u a lita tiv e v ariab le. can be conducted similarly.
T E S T IN G THE S I G N I F I C A N C E OF D U M M Y V A R IA B L E S
EXAMPLE 17.2
Refer to the regression results in Table 17.3.
a. Determine whether there is a difference in salary depending on gender at the
5% significance level.
b. Determine whether an older professor’s salary differs from a younger profes
sor’s salary at the 5% significance level.
SOLUTION:
a. In order to test for a salary difference between male and female professors,
we set up the hypotheses as H0: β 2 = 0 against HA: β 2 ≠ 0. Given a value of
the tdf test statistic of 4.86 with a p -value ≈ 0.00, we reject the null hypothesis
and conclude that the gender dummy variable is statistically significant at the
5% level. We conclude that male and female professors do not make the same
salary, holding other variables constant.
b. Here the hypotheses take the form H0: β 3 = 0 against HA: β3 ≠ 0. Given a
value of the tdf test statistic of 0.94 with a p -value = 0.36, we cannot reject the
null hypothesis. At the 5% significance level, we cannot conclude age discrimi
nation with respect to a professor’s salary. We should bear in mind that some of
the age effect is captured by experience. In the next section, we will consider a
scenario where males get a higher compensation for experience than females,
making their salaries diverge over time.
Assuming that the linear regression model includes an intercept, the number of
dummy variables representing a qualitative variable should be one less than the
number of categories of the variable.
EXAMPLE 17.3
A recent article suggests that Asian-Americans face serious discrimination in the
college admissions process (The Boston Globe, February 8, 2010). Specifically,
Asian applicants typically need an extra 140 points on the SAT to compete with
white students. Another report suggests that colleges are eager to recruit Hispanic
students who are generally underrepresented in applicant pools (USA Today,
February 8, 2010). In an attempt to corroborate these claims, a sociologist first
wants to determine if SAT scores differ by ethnic background. She collects data on
200 individuals from her city with their recent SAT scores and ethnic background.
A portion of the data is shown in Table 17.4; the complete data, labeled SAT and
Ethnicity, can be found on the text website.
S O LU TIO N :
2. Consider a linear regression model where y represents the a. Given that there are four quarters in a year, why doesn't
response variable, d 1is a dum m y variable assuming values of the analyst include a fourth dum m y variable in his
0 or 1, and d2 is another dum m y variable assuming values o f model?
0 or 1. The model is estimated as ŷ = 160 + 15d1 + 32d2. b. At the 5% significance level, are the dum m y variables
a. Compute ŷ for d1 = 1 and d2 = 1. individually significant? Explain. Is the analyst able to
b. Compute ŷ for d1= 0 and d2 = 0. obtain higher returns depending on the quarter?
3. Using 50 observations, the follow ing regression o u tput is 6. FILE In the United States, baseball has always been a
obtained from estimating y = β 0 + β1x + β 2d1 + β 3d2 + ε . favorite pastime and is rife w ith statistics and theories. While
baseball purists may disagree, to an applied statistician no
Standard
topic in baseball is too small or hypothesis too unlikely. In a
Coefficients Error t Stat p -value
recent paper, researchers at Wayne State University showed
Intercept – 0.61 0.23 – 2.75 0.0074
that major-league players w ho have nicknames live 21/2 years
x 3.12 1.04 3.01 0.0034
longer than those w ith ou t them (The Wall Street Journal,
d1 – 13.22 15.65 – 0.85 0.4006 July 16, 2009). Perhaps nicknames add to the self-esteem
d2 5.35 1.25 4.27 0.0000 of a player. Consider the follow ing portion o f data on the
lifespan (years) o f a player and a nickname dum m y that
a. Compute ŷ fo r x = 250, d1 = 1, and d2 = 0; then compute
equals 1 if the player had a nickname and 0 otherwise; the
ŷ fo r x = 250, d1 = 0, and d2 = 1.
complete data set can be found on the text website, labeled
b. Interpret d 1and d2. Are both dum m y variables
Nicknames.
individually significant at the 5% level? Explain.
Years Nickname
4. Using 30 observations, the follow ing regression output
74 1
is obtained from estimating ln(y) = β 0 + β1 x + β 2d + ε ,
where ln(y ) is the natural log o f y .
62 1
⋮ ⋮
Standard
Coefficients Error t Stat p-value 64 0
Intercept 1.56 0.73 2.14 0.0415 a. Create tw o subsamples, w ith one consisting of players
x 0.21 0.08 2.63 0.0139 w ith a nickname and the other one w ith ou t a nickname.
d 0.15 0.04 3.75 0.0008 Calculate the average longevity for each subsample.
The standard error of the estimate is se = 0.35. b. Estimate a linear regression model o f Years on the
Nickname dum m y variable. Compute the predicted
a. Interpret the estimated coefficient for the dum m y
longevity o f players w ith and w ith o u t a nickname.
variable d.
c. Conduct a one-tailed test at a 5% level to determine if
b. C om pute ŷ fo r x = 20 and d = 0, then compute ŷ for
players w ith a nickname live longer.
x = 2 0 a n d d = 1.
7. F I L E The SAT has gone through many revisions over the
c. Is d significant at the 5% level? Explain.
years. In 2005, a new w riting section was introduced that
includes a direct w riting measure in the form o f an essay.
Applications People argue tha t female students generally do worse on
5. In an a tte m p t to "tim e th e m arket," a fin a n cial analyst studies math tests but better on w riting tests. Therefore, the new
the quarterly returns o f a stock. He uses the model y = β 0 + section may help reduce the usual male lead on the overall
β 1 d 1 + β 2d2 + β 3d3 + ε where y is the quarterly return o f a average SAT score (The Washington Post, August 3 0 , 2006).
stock, d 1is a dum m y variable that equals 1 if quarter 1 a n d 0 Consider the follow ing portion o f data on 20 students who
otherwise, d2 is a dum m y variable tha t equals 1 if quarter 2 took the SAT test last year; the complete data set can be
found on the te xt website, labeled SATdummy. Inform ation (EXPER), age (AGE), and gender o f the em ployee (d equals
includes each student's score on the math and w riting 1 if male). A portion o f the data is shown; the entire data set
sections o f the exam. Also included are the student's GPA can be found on the te xt website, labeled Hourly Wage.
and a dum m y tha t equals 1 fo r fem ale and 0 for male. Wage EDUC EXPER AGE d
Writing Math GPA Female $37.85 11 2 40 1
620 600 3.44 0 21.72 4 1 39 0
570 550 3.04 0 ⋮ ⋮ ⋮ ⋮ ⋮
⋮ ⋮ ⋮ ⋮
24.18 8 11 64 0
540 520 2.84 0
a. Estimate: Wage = β 0 + β 1 EDUC + β 2EXPER + β 3AGE +
a. Estimate a linear regression m odel w ith w ritin g score as β 4d + ε .
the response variable and GPA and the fem ale dum m y
b. Predict the hourly wage o f a 40-year-old male
variable as the explanatory variables. em ployee w ith 10 years o f higher education and 5 years
b. Com pute the predicted score for a male student w ith a experience. Predict the hourly wage o f a 40-year-old
GPA o f 3.5. Repeat the analysis fo r a fem ale student. fem ale employee w ith the same qualifications.
c. Perform a test to determ ine if there is a statistically c. Interpret the gender coefficient. Is the gender variable
significant gender difference in w ritin g scores at a 5% significant at the 5% level? Does the data suggest tha t
level. gender discrim ination exists at this firm?
8. FILE Use the data described in Exercise 7 to estim ate a linear 11. F IL E A governm ent researcher is analyzing the relationship
regression model w ith m ath score as the response variable between retail sales and the gross national product (GNP).
and GPA and the fem ale dum m y variable as the explanatory He also wonders w hether there are significant differences
variables. in retail sales related to the quarters o f the year. He collects
a. Compute the predicted score fo r a male student w ith a ten years o f quarterly data. A portion is shown in the
GPA o f 3.5. Repeat the analysis fo r a fem ale student. accom panying table; the com plete data set can be found on
b. Perform a test to determ ine if there is a statistically the te xt website, labeled Retail Sales.
significant gender difference in math scores at the 5% Retail Sales GNP
level. Year Quarter (in $ millions) (in $ billions) d1 d2 d3
9. FILE A m anager at an ice cream store is try in g to determ ine 2001 1 696,048 9,740.5 1 0 0
how many customers to expect on any given day. Overall 2 753,211 9,983.5 0 1 0
business has been relatively steady over the past several ⋮ ⋮ ⋮ ⋮ ⋮ ⋮
years, b u t the custom er count seems to have ups and
2009 4 985,649 14,442.8 0 0 0
downs. He collects data over 30 days and records the
num ber o f customers, the high tem perature (degrees S o u r c e : Retail sales o b ta in e d fro m w w w .census.gov; GNP o bta ine d fro m research.
stlouisfed.org.
Fahrenheit), and w hether the day fell on a weekend
(1 equals weekend, 0 otherwise). A po rtio n o f the data a. Estimate y = β 0 + β1 x + β 2d 1 + β 3d2 + β 4d3 + ε where
is shown; th e entire data set can be found on the text y is retail sales, x is GNP, d 1 is a dum m y variable tha t
website, labeled Ice Cream. equals 1 if quarter 1 and 0 otherwise, d2 is a dum m y
Customers Temperature Weekend variable tha t equals 1 if quarter 2 and 0 otherwise, and
376 75 0 d3 is a dum m y variable th a t equals 1 if quarter 3 and 0
otherwise.
433 78 0
⋮ ⋮ ⋮ b. Predict retail sales in quarters 2 and 4 if GNP equals
$13,000 billion.
401 68 0
c. W hich o f the quarterly sales are significantly different
a. Estimate: Customers = β 0 + β1 Temperature + from those o f the 4th quarter at th e 5% level?
β 2Weekend + ε . d. Use the partial F test to determ ine if the three seasonal
b. How many customers should the manager expect on a dum m y variables used in the model are jo in tly significant
Sunday w ith a forecasted high tem perature o f 80°? at the 5% level.
c. Interpret the weekend coefficient. Is it significant at 12. FILE The issues regarding executive com pensation have
the 5% level? How m ig ht this affect the store's staffing received extensive media a tte n tio n . The governm ent is
needs? even considering a cap on high -flying salaries for executives
10. F IL E A researcher wonders w hether males get paid more, (New York Times, February 9 , 2009). Consider a regression
on average, than females at a large firm . She interview s model th a t links executive com pensation w ith the total
50 employees and collects data on each worker's hourly assets o f the firm and the firm 's industry. Dummy variables
wage (Wage), years o f higher education (EDUC), experience are used to represent four industries: M anufacturing
T echnology d 1 M anu facturing O th er d 2, Financial Services a. E stim ate th e m odel: ln(y ) = β 0 + β1 ln(x ) + β 2d 1 +
d 3, a n d Nonfinancial Services d 4. A p o rtio n o f th e d a ta for β 3d 2 + β 4d 3 + ε , w h e re ln(y ) a n d ln(x ) d e n o te th e log of
455 h ig h est-p a id CEOs in 2006 is given in th e accom panying c o m p e n sa tio n a n d th e log o f to ta l assets, respectively.
tab le; th e c o m p le te d ata, labeled In d u stry C om pen satio n , The nonfinancial services in du stry d u m m y d4 is u sed for
are available on th e te x t w ebsite. reference.
C om pensation A ssets b. In terp re t th e e stim a te d coefficients.
(in $ million) (in $ m illions) d1 d2 d3 d4 c. Use a 5% level o f significance to d e te rm in e w hich
16.58 20,917.5 1 0 0 0 industries, relative to th e nonfinancial services industry,
26.92 32,659.5 1 0 0 0 have a different executive co m p e n sa tio n .
⋮ ⋮ ⋮ ⋮ ⋮ ⋮ d. Use th e partial F te s t to d e te rm in e if th e th re e industry
T E S T I N G THE S I G N I F I C A N C E OF D U M M Y
A N D IN T E R A C T IO N VARIABLES
EXAMPLE 17.4
In Section 17.1 we estimated a regression model that tested for gender and age
discrimination in salaries. We found that the number of years of experience x and
the gender dummy d 1 were significant in explaining salary differences; however,
the age dummy d2 was insignificant. In an attempt to refine the model explaining
salary, we drop d2 and estimate three models using the data from Table 17.1, where
y represents annual salary (in $1,000s).
Model 1. y = β0 + β1 x + β 2d 1 + ε
Model 2. y = β0 + β1 x + β2 x d 1 + ε
Model 3. y = β0 + β1 x + β2 d 1 + β3 x d 1 + ε
SO LU TIO N :
a. In order to estimate the three models, we first generate data on both d 1and x d 1;
Table 17.6 shows a portion of the data.
TABLE 17.6 Generating d 1and xd1from the Data in Table 17.1
y x d1 x d 1
67.50 14 1 1 4 × 1 = 14
53.51 6 1 6 ×1 = 6
⋮ ⋮ ⋮ ⋮
73.06 35 0 35 × 0 = 0
N otes: The to p p ortion o f the tab le contains param eter estimates w ith p -values in parentheses; NA denotes n ot
applicable: * represents significance at th e 5% level; Adjusted R2, reported in th e last row, is used fo r m odel selection.
M odel 1 uses a gender dummy variable d 1 to allow salaries between males
and females to differ by a fixed amount, irrespective o f experience. It is
estim ated a s ŷ = 39.43 + 1.24x + 13.896d1. Since d 1 is associated with a
p -value o f 0.01, we conclude at the 5% level that d 1has a statistically sig
nificant influence on salary. The estim ated model implies that, on average,
males earn $13,890 (13.89 × 1,000) more than females at all levels of
experience.
Model 2 uses an interaction variable x d 1to allow the difference in salaries
between males and females to change with experience. It is estimated as
ŷ = 47.07 + 0.85x + 0 .77x d 1. Since x d 1is associated with a p -value ≈ 0.00,
we conclude that it is statistically significant at the 5% level. With every extra
year of experience, the estimated difference in salaries between males and
females increases by $770 (0.77 × 1,000).
Model 3 uses d 1 along with x d 1to allow a fixed as well as a changing dif
ference in salaries between males and females. The estim ated regression
equation i s ŷ = 49.42 + 0.76x – 4.00d1 + 0.93x d 1. Interestingly, the vari
able d 1 is no longer statistically significant at the 5% level with a p -value
of 0.42. However, the variable x d 1 is significant, suggesting that with every
extra year of experience, the estim ated difference in salaries between
males and females increases by $930 (0.93 × 1,000).
b. While Model 1 shows that the gender variable d 1is significant and Model 2
shows that the interaction variable x d 1 is significant, Model 3 provides some
what conflicting results. This raises an important question: which model
should we trust? It is not uncommon to contend with such scenarios in
business applications. As discussed in Chapter 14, we usually rely on adjusted
R2 to compare models that have a different number of explanatory variables.
Based on the adjusted R2 values of the models, reported in the last row of
Table 17.7, we select Model 2 as the preferred model because it has the highest
value of 0.7923.
c. In order to interpret the results further, we use Model 2 to estimate salaries
with varying levels of experience, for both males and females. For example,
with 10 years of experience, the predicted salary for males (d 1 = 1) is
ŷ = 47.07 + 0.85(10) + 0.77(10 × 1) = 63.27, or $63,270.
SYNOPSIS OF I N T R O D U C T O R Y CASE
d is a d u m m y variable a ssu m in g values 0 or 1. A m odel w ith p red ict co n su m p tio n for Incom e = $75,000 in urban
a. C o m p u te ŷ f o r x = 1 0 a n d d = 1. Explain.
A LINEAR P RO BA BILITY M O D E L
EXAMPLE 17.5
The subprime mortgage crisis has forced financial institutions to be extra stringent in
granting mortgage loans. Many seemingly creditworthy applicants are having their loan
requests denied. Thirty recent mortgage applications were obtained to analyze the mort
gage approval rate. The response variable y equals 1 if the mortgage loan is approved,
0 otherwise. It is believed that approval depends on the percentage of the down pay
ment x 1 and the percentage of income-to-loan amount x2. Table 17.9 shows a portion
of the data; the complete data set can be found on the text website, labeled Mortgage
Applications. Estimate and interpret the linear probability model, y = β 0 + β1x1 +
β2x2 + ε . Make predictions of the approval probability for representative applicants.
SOLUTION: Table 17.10 shows the relevant regression results. The estimated
regression equation is Note that both
explanatory variables exert a positive and statistically significant influence on loan
approval at a 5% level, with p -values of 0.0120 and 0.0003, respectively. Also,
b1 = 0.0188 implies that a 1-percentage-point increase in down payment increases
the approval probability by 0.0188, or by 1.88%. Similarly, a 1-percentage-point
increase in the income-to-loan ratio increases the approval probability by 0.0258.
We can use this estimated model to predict the approval probability for any appli
cant. Consider an applicant who puts 20% down (x 1 = 20), has an income of $60,000
and a loan amount of $200,000 (x 2 = (60/200) × 100 = 30). The predicted ap
proval probability for this applicant is
Similarly, with 30% down,
In other words, as down payment increases by 10 percentage points, the
predicted probability of approval increases by 0.1880 (= 0.4698 – 0.2818), which
is essentially the estimated slope, 0.0188, multiplied by 10. The estimated slope
coefficient for the percentage of income-to-loan variable can be analyzed similarly.
Although it is easy to estimate and interpret a linear probability model, it also has some
shortcomings. The major shortcoming is that it can produce predicted probabilities that
are greater than 1 or less than 0. For instance, for a down payment of 60%, with the
same income-to-loan ratio of 30%, we get a predicted mortgage approval rate of
a probability greater than one! Similarly,
for a down payment of 5%, the model predicts a negative probability,
Furthermore, the linearity of the relationship may
also be questionable. For instance, we would expect a big increase in the probability of
loan approval if the applicant makes a down payment of 30% instead of 20%. This increase
in probability is likely to be much smaller if the same 10-percentage-point increase in
down payment is from 60 to 70 percent. An LPM cannot differentiate between these
two scenarios. For these reasons, we introduce a more appropriate probability model for
binary choice variables.
The logit model is estimated with the method of maximum likelihood estimation
(MLE). Predictions with this model are made by
where b0 and b1 are the MLE estimates of the population parameters β0 and β 1 .
Figure 17.4 highlights the relationship between the predicted probability and the
explanatory variable x for an LPM and a logit probability model, given b1 > 0. Note
that in an LPM, the probability falls below 0 for small values of x and exceeds 1 for
large values of The probabilities implied by a logit model, however, are always
constrained in the [0,1] interval. (For ease of exposition, we use the same notation
to refer to the coefficients in the LPM and logit model. We note, however, that these
coefficients and their estimates have a different meaning depending on which model
we are referencing.)
FIGURE 17.4 Predicted probabilities w ith an LPM and
a lo git model
EXAMPLE 17.6
There is a declining interest among teenagers in pursuing a career in science and
health care (US News and World Report, May 23, 2011). In a recent survey, 50%
of high school students showed no interest in the sciences. An educator wants to
determine if a student’s interest in the sciences is linked with the student’s GPA.
She estimates a logit model where the choice of field (1 for choosing science, 0
otherwise) depends on the student’s GPA. She uses Minitab to produce the logit
regression results shown in Table 17.11. (Instructions for Minitab as well as other
software packages can be found on the text website.)
Similarly, we compute the predicted probabilities for a student with GPA = 3.5
and GPA = 4.0 as 0.72 and 0.84, respectively. Note that the predicted proba
bility increases by 0.18 (= 0.72 – 0.54) as GPA increases from 3.0 to 3.5. The
increase is only 0.12 ( = 0.84 – 0.72) when GPA increases from 3.5 and 4.0.
IE
EXAMPLE
F
L 17.7
Let us revisit Example 17.5, based on the Mortgage Applications data, a portion of
which was presented in Table 17.9. Estimate and interpret a logit model for mort
gage approval y based on the applicant’s percentage of down payment x 1 and the
applicant’s percentage of income-to-loan ratio x2. Make predictions of the approval
probability for representative applicants. Compare the results of the logit model
with those of the LPM.
S O LU TIO N : We again use Minitab to estimate the logit model; Table 17.12 shows
a portion of the output.
As in the case of the linear probability model, both variables exert a positive and
statistically significant influence on loan approval at a 5% level, with positive esti
mated coefficients and p -values of 0.0351 and 0.0058, respectively.
We can use the estimated model to predict approval probabilities for any
applicant. For instance, for an individual with x 1 = 20 and x2 = 30, the predicted
approval probability is
As discussed earlier, with a linear probability model, the predicted probabilities can
be negative or greater than one. The probabilities based on a logit model always stay
between zero and one for all possible values of the explanatory variables. Therefore,
whenever possible, it is preferable to use the logit model over the LPM for binary
choice models.
EXERCISES 17.3
Variable LPM Logit 25. Using 30 observations, the fo llow ing o u tp u t was obtained
when estim ating the lo g it model.
– 0.40 – 4.50
Constant Predictor Coef SE Z P
(0.03) (0.01)
0.32 1.54 Constant – 0.188 0.083 2.27 0.024
x
(0.04) (0.03) x 3.852 1.771 2.18 0.030
a. Use both models to predict th e pro ba bility o f success as a. W hat is the predicted pro ba bility when x = 0.40?
x varies from 1 to 5 w ith increm ents o f 1. b. Is x significant at the 5% level?
26. Using 40 observations, the follow ing o u tput was obtained and th e individual's incom e (in $1,000s). The com plete data set
when estimating the logit model. can be found on the text website, labeled Health Insurance.
Predictor Coef SE Z P Premium
Constant 1.609 1.405 1.145 0.252 Insurance Percentage (in %) Income (in $1,000s)
x1 – 0.194 0.143 – 1.357 0.177 1 0 88
x2 0.202 0.215 0.940 0.348 0 0 60
x3 0.223 0.086 2.593 0.010 ⋮ ⋮ ⋮
a. What is the predicted probability when x1 = 15, x2 = 10, 0 60 60
a n d x 3 = – 2? a. Analyze a linear probability model for insurance
b. At the 5% significance level, which o f the explanatory coverage w ith premium percentage and income used as
variables are significant? the explanatory variables.
b. Consider an individual w ith an income o f $60,000. What
Applications
is the probability tha t she has insurance coverage if her
27. F I L E Annabel, a retail analyst, has been follow ing Under
employer contributes 50% o f the premium? What if the
Armour, Inc., the pioneer in the compression-gear market.
employer contributes 75% o f the premium?
Compression garments are meant to keep moisture away
30. FILE Analyze a logit model w ith the above Health Insurance
from a wearer's body during athletic activities in warm and
data. Consider an individual w ith an income o f $60,000.
c o o l w e a th e r. A n n a b e l b e lie v e s th a t th e U n d e r A rm o u r b ra n d
W hat is the probability that she has insurance coverage if
attracts a younger customer, whereas the more established
her employer contributes 50% o f the premium? What if the
companies, Nike and Adidas, draw an older clientele. In
em ployer contributes 75% o f the premium?
order to test her belief, she collects data on the age of
the customers and whether or not they purchased Under 31. F IL E According to a recent estimate, th e divorce rate in England
Arm our (1 for Under Armour, 0 otherwise). A portion o f the has fallen to a 26-year low (The Guardian, August 2 9 , 2008).
data is shown in the accompanying table; the complete data However, it is docum ented th a t th e rate o f divorce is m ore than
set is found on the text website, labeled Purchase. tw ice as high fo r men and w o m en aged 25 to 29. John Haddock
is a sociologist from Sussex University w h o w ants to analyze th e
Under Armour Age
divorce rate based on th e individual's age, fam ily income, and
1 30
th e n u m b e r o f children th a t th e couple has. He collects data on
0 19
⋮ ⋮ 30 individuals in a small to w n near Brighton, a p o rtio n o f which
is shown in th e accom panying table; th e com plete data set can
1 24
be fo u nd on th e te xt website, labeled Divorce.
a. Estimate a linear p ro b a b ility m odel using Under A rm ou r as
Divorce Age Income (in £1,000s) Children
th e response variable and age as th e explanatory variable.
0 1 19 3
b. Compute the predicted probability o f an Under Armour
0 0 46 3
purchase for a 20-year-old customer and a 30-year-old ⋮ ⋮ ⋮ ⋮
customer.
0 0 26 0
c. Test Annabel's belief that the Under Arm our brand
attracts a younger customer at the 5% level. a. Estimate and interpret a linear probability model
where divorce (1 for divorce; 0 otherwise) depends on
28. F ILE Use the above data labeled Purchase to estimate a
age (1 if 25– 29 years old; 0 otherwise), fam ily income
lo git model.
(in £1,000s), and the num ber o f children.
a. C o m p u te th e pre d icte d p ro b a b ility o f an U nde r A rm o u r p u r
chase fo r a 20 -year-old c u sto m e r a n d a 30 -year-old custom er.
b. Do the data support the article's claim that the divorce
rate is higher for those aged 25– 29 years old? Explain.
b. Test Annabel's belief tha t the Under Arm our brand
c. Use the above estimates to predict the probability
attracts a younger customer at the 5% level.
o f divorce for an individual w ho is 27 years old, has
29. F IL E A c c o rd in g to th e N a tio n a l C o a litio n o n H e a lth Care, th e re
£60,000 o f fam ily income and one child. Recalculate the
has b e e n a ste a d y d e c lin e in th e p ro p o r tio n o f A m e rica n s w h o
probability w ith three children.
have health insurance.The rising insurance premiums have
32. F I L E Estimate and interpret the lo git model w ith the above
made it difficult for small employers to offer insurance and
Divorce data.
those th a t do o ffer insurance are co n trib u tin g a sm aller share
o f th e p rem ium . As a result, an increasing n u m ber o f Americans a. Do the data support the article's claim that the divorce
do not have health insurance because they cannot afford it. rate is higher for those aged 25–29 years old? Explain.
Consider a portion o f data in the following table relating to b. Use the above estimates to predict the probability
insurance coverage (1 for coverage, 0 for no coverage) for 30 o f divorce for an individual w ho is 27 years old, has
working individuals in Atlanta, Georgia. Also included in the £60,000 o f fa m ily incom e, and one child. Recalculate th e
table is the percentage of the premium paid by the employer probability w ith three children.
WRITING WITH STATISTICS
During the 2009-2010 NBA season, the Los Angeles Lakers had
the highest offensive production throughout the league. Led by
Kobe Bryant, the Lakers beat the Boston Celtics in game seven of
the championships for the 2010 NBA title. Jaqueline Thomsen, an
amateur statistician, would like to examine the factors that led to the
Lakers’ success. Specifically, Jaqueline wishes to predict the likeli
hood of a Lakers’ win as a function of field goal percentage (FG),
rebounds (Rebounds), and turnovers (Turnovers). The probability of
winning should be positively influenced by FG and Rebounds, but
negatively affected by Turnovers. In addition, she wonders whether
playing at home significantly influences the team’s chances of win
ning. Table 17.14 shows a portion of data on the Lakers’ 82-game
regular season; the complete data set can be found on the text web
site, labeled Lakers.
Source: w w w .n b a .c o m .
With the highest offensive production throughout the league during the 2009-2010 Sample
season, it is not surprising that the Los Angeles Lakers won the 2010 NBA champion
ship. Other teams might benefit if they could unravel the factors that led to the Lak Report—
ers’ success. In an attempt to examine the factors that influence a team ’s chances of
winning, regression analysis is conducted on the Lakers’ 82-game regular season. The Predicting the
response variable is Win (equals 1 for a win, 0 otherwise) and the explanatory variables
include: Probability
• The team ’s field goal percentage (FG),
o f Winning
• The number of rebounds,
• The number of turnovers, and
• A “home” dummy that equals 1 for a home game and 0 otherwise.
The probability of winning should be positively influenced by FG and rebounds, but
negatively affected by turnovers. In addition, if there truly is a home court advantage, then
playing at home should positively influence the team’s chances of winning.
Two models are evaluated that link the probability of winning with the explanatory
variables: the linear probability model and the logit model. The parameter estimates of
both models are shown in Table 17.A.
TABLE 17.A Regression Results o f the Linear P robability Model and the Logit Model
Response Variable: Win (equals 1 if Lakers win, 0 otherwise)
LPM Logit
– 2.391* – 28.76*
Constant
(0.00) (0.00)
0.047* 0.49*
FG
(0.00) (0.00)
0.019* 0.17*
Rebounds
(0.00) (0.02)
– 0.004 – 0.04
Turnovers
(0.68) (0.71)
0.232* 1.82*
Home
(0.01) (0.02)
5 % le ve l.
As in the case of the linear probability model, FG, Rebounds, and Home are again indi
vidually significant at the 5% level; thus, the significance of Home supports the belief of
a home field advantage. Over the 82-game season, the averages for field goal percentage,
the number of rebounds, and the number of turnovers were 45%, 44, and 13, respectively.
If the Lakers are playing an “average” game away from home, then the model predicts a
56.2% probability of winning. However, if they are playing an “average” game at home,
then their probability of winning jumps to 88.8%. In sum, the home court advantage over
whelmingly puts the likelihood of success in their favor.
Conceptual Review
LO 17.1 Use dummy variables to capture a shift of the intercept.
A dummy variable d is defined as a variable that takes on values of 0 or 1. It is used
to represent two (or more) categories of a qualitative variable. The number of dummy
variables for a multi-category qualitative variable should be one less than the number
of categories of the variable.
A regression model with a dummy variable d, representing two categories of a quali
tative variable, and a quantitative variable x is specified by y = β0 + β1x + β 2d + ε .
We estimate this model to make predictions a s ŷ = (b0 + b2) + b 1x for d = 1 and as
ŷ = b0 + b 1x for d = 0.
LO 17.4 U s e a lin e a r p r o b a b il it y m o d e l t o e s t i m a t e a b in a r y r e s p o n s e v a r ia b le .
Models that use a dummy (binary) variable as the response variable are called binary
choice models. A linear probability model (LPM) is formulated as y = β 0 + β1x1 +
β2x2 + . . . + βkxk + ε = P(y = 1) + ε , where y assumes values of 0 or 1 and P(y = 1)
is the probability of success.
Predictions with this model are made by where
b0, b 1, b2, . . . , bk are the estimates.
The major shortcoming of the LPM is that it can produce predicted probabilities that
are greater than one or less than zero.
Year Quarter Return d1 d2 d3 34. FILE In a sem inal study, researchers d o cu m en ted race-b ased
hiring in th e Boston an d Chicago labor m arkets (American
2000 1 4.85 1 0 0
Economic Review, S ep tem b er 2004). They se n t o u t identical
2000 2 – 3.96 0 1 0
resum es to em ployers, h alf w ith traditionally African-Am erican
⋮ ⋮ ⋮ ⋮ ⋮ ⋮
nam es and th e o th e r half w ith traditionally C aucasian nam es.
2009 4 4.06 0 0 0 Interestingly, th e re w as a 53% difference in call-back rates
S ource: h t t p : / /f i n a n c e . y a h o o . c o m . b etw een th e tw o gro up s o f people. A research fellow at
an institute in Santa Barbara decides to repeat th e sam e a. E stim ate th e m odel, BMI = β 0 + β 1 Fem ale + β 2Black
e x p e r im e n t w ith n a m e s a lo n g w ith a g e in t h e Los A n g eles lab o r + β3(Fem ale × Black) + ε , to p red ict th e BMI for
m a rk e t. S h e re p e a te d ly s e n d s o u t re s u m e s fo r sales p o sitio n s in w h ite m ales, w h ite fem ales, black m ales, a n d black
t h e city t h a t a re id en tical e x c e p t for t h e d ifferen ce in t h e n a m e s fem ales.
a n d a g e s o f t h e a p p lican ts. S h e also reco rd s t h e call-b ack rate b. Is th e difference b e tw e e n w h ite fem ales a n d w h ite m ales
for e a c h c a n d id a te . T h e a c c o m p a n y in g ta b le s h o w s a p o rtio n statistically significant a t th e 5% level?
o f d a ta o n call-b ack ra te (%), a g e , a n d a C au casian d u m m y th a t c. Is th e difference b e tw e e n w h ite m ales a n d black m ales
e q u a ls 1 fo r a C a u c a sia n -so u n d in g n a m e ; t h e c o m p le te d a ta se t statistically significant a t th e 5% level?
c a n b e f o u n d o n t h e te x t w e b s ite , la b e le d Hiring.
37. F I L E A ccording to th e C en ter for D isease Control a n d
Call-back Age Caucasian Prevention, life e x p ec ta n c y a t a g e 65 in Am erica is a b o u t
12 60 1 18.7 years. Medical research ers have a rg u e d th a t w hile
9 56 0 excessive drinking is d e trim e n ta l to h ealth, drinking a little
⋮ ⋮ ⋮ alcohol every day, especially w ine, m ay b e a sso ciated w ith an
in crease in life exp ectancy. O thers have also linked longevity
15 38 0
w ith in co m e a n d g e n d er. The a cco m p an y in g ta b le sh ow s a
a. E stim ate a linear regression m o del w ith call-back as th e p o rtio n o f d a ta relating to th e len g th of life a fte r 65, av era g e
re sp o n se variable, a n d a g e a n d th e Caucasian d u m m y as in co m e (in $1,000s) a t a re tire m e n t a g e of 65, a "w om an"
th e ex p la n ato ry variables. dum m y, a n d th e av era g e n u m b e r o f alcoholic drinks
b. C o m p u te th e call-back rate for a 30-year-old a p p lic an t c o n su m e d p e r day. The full d a ta se t can b e fo u n d on th e tex t
w ith a C au casian -so und in g na m e . W hat is th e w ebsite, labeled L on gevity.
co rre sp o n d in g call-back rate for a non-C aucasian? Life Incom e (in $1,000) Woman Drinks
c. C o n d u c t a te s t for race d iscrim ination a t th e 5% 19.00 64 0 1
significance level. 19.30 43 1 3
35. An a n alyst stu d ies q u a rte rly d a ta o n th e relationship ⋮ ⋮ ⋮ ⋮
b e tw e e n retail sales (y , in $ millions), gro ss natio nal p ro d u c t 20.24 36 1 0
(x, in $ billions), a n d a q u a rte rly d u m m y d th a t e q u als 1 if th e
sales are for th e 4 th q u a rte r; 0 o th erw ise. He e stim a te s th e a. Use th e d a ta to m od el life e x p ec ta n c y a t 65 on th e basis
m o del y = β 0 + β1x + β 2d + β3x d + ε . Relevant regression of Incom e, W om an, a n d Drinks.
a. In terp re t th e in te rce p t dum m y. Is it significant a t th e 5% Like m o st universities, it u ses SAT scores a n d high school
GPA as p rim ary criteria for adm ission . The accom p any ing
level?
ta b le sho w s a p ortion o f d a ta c o n ce rn in g inform ation on
b. In terp re t th e in teractio n variable. Is it significant a t th e
a d m issio n (1 for adm issio n a n d 0 otherw ise), SAT score, a n d
5% level?
GPA for 30 stu d e n ts w h o h a d recently a p p lied to S eton Hall;
36. F I L E A ccording to th e U.S. D e p a rtm e n t of H ealth a n d H um an
th e full d a ta s e t can b e fo u n d o n th e te x t w ebsite, labeled
Services, African-Am erican w o m e n have th e h ig h e st rates
S e to n Hall.
o f b e in g o v e rw eig h t c o m p a red to o th e r g ro u p s in th e U.S.
Adm ission SAT GPA
Individuals a re co n sid ered o v e rw eig h t if th e ir bod y m ass
index (BMI) is 25 or greater. C onsider th e follow ing d a ta on 1 1700 3.39
BMI o f 120 individuals a n d th e co rre sp o n d in g g e n d e r and 1 2020 2.65
race du m m ie s. The c o m p le te d a ta , labeled O v erw eig h t, can ⋮ ⋮ ⋮
be fo u n d on th e te x t w eb site. 0 1300 2.47
re-o ffen d . A so cio lo g ist collects d a ta o n 2 0 individuals w h o w ere 41. F IL E E s tim ate th e logit m odel w ith th e abo ve P arole d ata.
released o n p aro le tw o years ag o . S he n o te s if h e /s h e c o m m itte d a. Are th e results c o n sisten t w ith th e claims o f o th e r stu d ies
a n o th e r crim e o v er t h e last tw o years (crim e eq u a ls 1 if crim e w ith re sp ec t to a g e a n d g en d er?
co m m itte d , 0 o th erw ise), th e individual's a g e a t t h e tim e of b. Predict th e probability o f a 25-year-old m ale p arolee
release, a n d t h e g e n d e r o f th e individual (g e n d e r e q u a ls 1 if m ale, c o m m ittin g a n o th e r crim e; re p e a t th e prediction for a
0 o th erw ise). T he a c c o m p a n y in g ta b le s h o w s a p o rtio n o f d ata; 25-year-old fem ale parolee.
t h e full d a ta s e t can b e fo u n d o n t h e te x t w eb site, lab eled Parole.
CASE STUDIES
LEARNING OBJECTIVES
R
Time Series and
Forecasting
After reading LO 18.1 Distinguish among the various models used in forecasting.
this chapter
you should LO 18.2 Use smoothing techniques to make forecasts.
be able to:
LO 18.3 Use trend regression models to make forecasts.
LO 1 8.4 Calculate and interpret seasonal indices and use them to seasonally adjust
a time series.
LO 18.6 Use trend regression models with seasonal dummy variables to make
forecasts.
Forecasting is an im p o rtan t aspect o f statistical analysis, providing guidance for decisions in all areas
o f business. Examples are abundant and include forecasting product sales, the inflation rate, the
price o f a financial asset, or a company's cash flows. In fact, the success o f any business or govern
m ent agency depends on the ability to accurately forecast many vital variables. Sound forecasts not
only improve the quality o f business plans, but also help identify and evaluate potential risks. The
field o f forecasting has developed rapidly over the last few decades, w ith some approaches requir
ing highly sophisticated techniques. In this chapter we focus on some o f the easier approaches,
w hich nevertheless provide a flavor and insight into this fascinating field. In particular, we use simple
sm oothing techniques for making forecasts when short-term fluctuations in the data represent ran
d om departures from the overall pattern w ith no discernible trend or seasonal fluctuations. Special
forecasting methods are introduced w hen trend and seasonal fluctuations are present in the data.
We will also explain a regression approach for forecasting.
I N T R O D U C T O R Y C A S E
1. D e te rm in e w h e th e r re v e n u e e x h ib its a n y s o rt o f tre n d .
Forecasting Methods
Forecasting methods are broadly classified as quantitative or qualitative. Qualitative
forecasting procedures are based on the judgment of the forecaster, who uses prior expe
rience and expertise to make forecasts. On the other hand, quantitative forecasting uses a
formal model along with historical data for the variable of interest.
Qualitative forecasting is especially attractive when past data are either not avail
able or are misleading. For instance, a manager may use qualitative forecasts when she
attempts to project sales for a brand new product, or when a major structural change in
market conditions has rendered previous data obsolete. Similarly, an economist may use
qualitative forecasts of credit flow resulting from a newly introduced stimulus package by
the federal government.
Although attractive in certain scenarios, qualitative forecasts are often criticized on
the ground that they are prone to some well-documented biases such as optimism and
overconfidence. Decisions based on the judgment of an overly optimistic manager may
prove costly to the business. Also, qualitative forecasting is difficult to document and its
quality is totally dependent on the judgment and skill of the forecaster. Two people with
access to similar information may offer different qualitative forecasts.
In this chapter we focus on quantitative forecasting. Formal quantitative models have
been used extensively to forecast variables such as sales, inflation, and housing starts.
These models are further split up into causal and noncausal models. Causal methods are
based on a regression framework, where the variable of interest is related to a single or
multiple explanatory variables. In other words, forecasts are “caused” by the known val
ues of the explanatory variables. Noncausal models, also referred to as purely time series
models, do not present any explanation of the mechanism generating the variable of inter
est and simply provide a method for projecting historical data. Despite the lack of theory,
noncasual models can provide sound forecasts. However, they provide no guidance on
the likely effects of changes in policy (explanatory) variables. Both types of quantitative
forecasting methods are discussed in this chapter, although the emphasis is on noncausal
methods.
TYPES OF F O R E C A S T IN G M E T H O D S
The in-sample forecast ŷt is also called the predicted or fitted value of yt. As always,
the residuals are computed as et = yt – ŷ t .
M O D E L SELECTION CRITERIA
The mean square error (M SE) and the mean absolute deviation (M A D ) are
computed as
where n is the number of residuals used in the computation. We choose the model
with the lowest M SE and MAD.
In the following sections we employ various forecasting models and compare them on the
basis of these goodness-of-fit measures.
1Here the MSE form ula is different from the one defined in Chapter 14, where the error sum o f squares was divided
by the appropriate degrees of freedom.
18.2 Smoothing Techniques
LO 18.2 Time series generally consist of systematic and unsystematic patterns. Systematic pat
Use smoothing terns are caused by a set of identifiable components, whereas unsystematic patterns by
techniques to definition are difficult to identify. Three identifiable components occur in systematic pat
make forecasts. terns: the trend, the seasonal, and the cyclical components. In this section we focus on
applications where the time series is described primarily by unsystematic patterns. In the
following sections we discuss systematic patterns.
Unsystematic patterns are caused by the presence of a random (irregular) error
term. As mentioned earlier, a time series is a sequence of observations that are ordered
in time. Inherently, any data collected over time is likely to exhibit some form of random
variation. For instance, the checkout time at a campus bookstore or weekly sales at a con
venience store encounter random variations for no apparent reason.
T I M E SERIES PA T T ERNS
A simple plot of the time series provides insights into its components. A jagged appear
ance, caused by abrupt changes in the series, indicates random variations. Smoothing
techniques are often employed to reduce the effect of random fluctuations. These tech
niques can also be used to provide forecasts if short-term fluctuations represent random
departures from the structure, with no discernible systematic patterns. These techniques
are especially attractive when forecasts of multiple variables need to be updated fre
quently. For example, consider a manager of a convenience store who has to update the
inventories of numerous items on a weekly basis. It is not practical in such situations to
develop complex forecasting models. We discuss two distinct smoothing techniques: the
moving average and the exponential smoothing techniques.
CALCULATING A M O V IN G AVERAGE
An m -period moving average is computed as
EXAMPLE 18.1
According to the Energy Information Administration, the United States consumes
about 21 million barrels (882 million gallons) of petroleum each day. About half of
this consumption is in the form of gasoline. Table 18.2 shows a portion of weekly
U.S. finished motor gasoline production, measured in thousands o f barrels per day;
the fu ll data set can be found on the text website, labeled Gas Production.
a. Construct a 3-period moving average series fo r the data.
b. Plot production and its corresponding 3-period moving average and comment
on any differences.
c. Using the 3-period moving average series, forecast gasoline production on
May 29, 2009 (week 22).
d. Calculate the mean square error M SE and the mean absolute deviation MAD.
SOLUTION:
a. For notational sim plicity, let production be denoted by yt and the correspond
ing moving average be denoted by ȳt. We form a 3-period moving average
series by averaging all sets o f three consecutive values o f the original series.
The first value o f a 3-period m oving average is calculated as
Other values o f ȳt are calculated sim ilarly and are presented in column 3 o f
Table 18.3. Note that we lose one observation at the beginning and one at the
end o f the 3-period moving average series ȳt. ( If it were a 5-period moving
average, we would lose two observations at the beginning and two at the end.)
y ŷ
e
| e |
(1) (2) (3 ) (4 ) (5) (6 )
7)(
1 9,115 — — — — —
2 8,813 8,885.67 — — — —
—
3 8,729 8,734.00 — — —
4 8,660 8,689.33 8,885.67 – 225.67 50,925.44 225.67
5 8,679 8,610.33 8,734.00 – 55.00 3,025.00 55.00
⋮ ⋮ ⋮ ⋮ ⋮ ⋮ ⋮
19 8710 8787.67 8932.00 – 222.00 49,284.00 222.00
20 8,735 8,941.00 8,806.00 – 71.00 5,041.00 71.00
21 9,378 — 8,787.67 590.33 348,493.44 590.33
Total 953,509.78 3,312.67
b. In Figure 18.1, we plot production and its corresponding 3-period moving
average against weeks. Note that the original production series has a jagged
appearance, suggesting the presence o f an important random component o f
the series. The series o f m oving averages, on the other hand, presents a much
smoother picture.
c. As mentioned earlier, if the prim ary component o f the series is random varia
tions, we can use moving averages to generate forecasts. Since ȳ2 represents
the average in weeks 1 through 3, it is the most updated estimate o f the series
prior to period 4. Therefore, w ith a 3-period moving average, ŷ4 = ȳ2 where ŷ4 is
the in-sample forecast fo r period 4. S im ilarly, ŷ5 = ȳ3 is the forecast fo r period
5, where ȳ3 is the average in weeks 2 through 4, and so on. These forecasts,
derived as are shown in column 4 o f Table 18.3. Follow ing
this simple process, we compute the out-of-sample forecast in week 22 as
Therefore, our forecast fo r gasoline production on May 29, 2009 (week 22) is
8,941 thousand barrels. One potential weakness when using the moving average
technique is that all future forecasts take on the same value as the first out-of-
sample forecast; that is, the forecast fo r week 23 is also 8,941 thousand barrels.
d. In order to calculate the mean square error M SE and the mean absolute devia
tion MAD, we first compute the residuals as et = y, – ŷt , shown in column 5 o f
Table 18.3. These residuals are squared (see column 6) and then summed to
compute MSE as
The absolute values o f the residuals, presented in column 7, are used to compute
MAD as
C A L C U L A T I N G A N E X P O N E N T I A L L Y S M O O T H E D SERIES
The exponential smoothing procedure continually updates the level of the series as
A t = α yt + (1 – α )A t–1,
where α represents the speed of decline. Forecasts are made as ŷt+1 = A t.
In order to implement this method, we need to determine α and the initial value of the
series A 1. Typically, the initial value is set equal to the first value of the time series, that
is, A 1 = y 1; the choice of the initial value is less important if the number of observations
is large. The optimal value for a is determined by a trial-and-error method. We evaluate
various values of α and choose the one that results in the smallest M SE and MAD.
SO LU TIO N :
c. Forecasts given b y ŷt+1 = A t are presented in column 4 of Table 18.4. For in
stance, for period 2, ŷ2 = A 1 = 9,115. Similarly, A 2 = 9,054.60 is the forecast
for ŷ3. Therefore, the forecast for gasoline production on May 29, 2009, com
puted as ŷ 22 – A 21, equals 8,933.29 thousand barrels. As with the moving aver
age method, any further out-of-sample forecasts also assume this same value;
for instance, ŷ 23 = 8,933.29 thousand barrels.
d. In columns 5, 6, and 7 we present the residuals, their squares, and their abso
lute values, respectively. We compute model selection measures as
Here, the choice o f α depends on whether we employ MSE or MAD for compari
son w ith MSE suggesting α = 0.4 and MAD suggesting α = 0.8. In instances
where MSE and MAD give conflicting results, we choose the procedure w ith the
smallest MSE; we make this choice because MSE penalizes larger deviations more
harshly due to squaring. Therefore, we choose α equal to 0.4 since it has the small
est MSE o f 53,364. In this application, the moving average model still outperforms
the exponential smoothing model, as measured by its lower MSE and MAD.
EXERCISES 18.2
c. Use a 5-period moving average to make a forecast for for December 1,2010.
the share o f rock music in 2009. b. Use the exponential smoothing method to make a
d. Use the MSE to pick the appropriate m oving average price forecast for December 1,2010. Use α = 0.4.
for making a forecast for the share o f rock'n'roll music c. Which o f the above sm oothing methods results in a
in 2009. lower MSE?
6. F I L E Use the data from the preceding question for the share d. You find out tha t the actual S&P 500 closing price
o f total shipm ent o f music tha t falls in the category o f rock on December 1,2010 was 1,206.07. Was the forecast
b. Make a forecast for the share o f rock music in 2009 using data on seasonally adjusted inflation and unem ploym ent
the exponential sm oothing m ethod w ith α = 0.6. rates in the United States from January 2009 to November
2010; the entire data set is on the text website, labeled a. Use a 3-period moving average and exponential
Unemployment and Inflation. sm oothing w ith α = 0.6 to make in-sample forecasts for
unem ploym ent. Use the more appropriate smoothing
Year Month Unemployment Inflation
m ethod to forecast unem ploym ent for December 2010.
2009 Jan 7.7 0.3
b. Use a 3-period moving average and exponential
⋮ ⋮ ⋮ ⋮
sm oothing w ith α = 0.6 to make in-sample forecasts for
2010 Nov 9.8 0.1 inflation. Use the more appropriate smoothing method
S o u r c e : Bureau o f Labor Statistics. to forecast inflation for December 2010.
EXAMPLE 18.3
The United States continues to increase diversity, with more than a third of its popu
lation belonging to a minority group (CNN.com, May 14, 2009). Hispanics are the
fastest-growing minority segment, comprising one out of six residents in the coun
try. Table 18.6 shows a portion of data relating to the number as well as the median
income of Hispanic households from 1975 through 2007. The full data set is on the
text website, labeled Hispanic Characteristics.
a. Use the sample data to estimate the linear trend model for the number (regres
sion 1) and the median income (regression 2) of Hispanic households. Interpret
the slope coefficients.
b. Forecast the number and the median income of Hispanic households in 2008.
S O LU TIO N : In order to estimate the above trend models, we first relabel the 33 years
of observations from 1 to 33. In other words, we make the explanatory variable t
assume values 1, 2, . . . , 33 rather than 1975, 1976, . . . , 2007. Table 18.7 uses a
portion of the data to show how t is created.
Relevant portions of the Excel output for the linear regression models are presented
in Table 18.8.
NOTES: Parameter estimates are follow ed w ith th e p -values in parentheses; *represents significance at 5% .
It is always advisable to inspect the data visually as a first step. Graphs offer an informal
way to gauge whether a linear or an exponential trend provides a better fit. Figures 18.4
and 18.5 are scatterplots of the number and the median income of Hispanic households
from 1975 through 2007. We relabel the 33 years of annual observations from 1 to 33 and
also superimpose the linear and the exponential trends to the data.
TABLE 18.9 G enerating the Natural Log o f the Series (Example 18.4)
Year t Number Income In (Number) In (Income)
1975 1 2,948 $8,865 7.9889 9.0899
1976 3,081 $9,569 8.0330 9.1663
2
⋮ ⋮ ⋮ ⋮ ⋮ ⋮
2007 33 13,339 38,679 9.4984 10.5631
Relevant portions of the Excel output for the exponential regression models are
presented in Table 18.10.
Se 0.0311 0.0775
N otes: Parameter estim ates are follow ed w ith th e p -values in parentheses; *represents significance at th e 5%
level. The last row shows th e standard error o f th e estim ate Se.
a. Consistent with the results of the linear trend models, the exponential trend
models suggest that the number and the median income of Hispanic house
holds are trending upward, since both regressions show positive slope coef
ficients: 0.0479 for regression 1 and 0.0418 for regression 2. In addition, the
slope coefficients are highly significant, with the p -values approximately zero
in both regressions.
b. We make forecasts as ŷ t = exp(b0 + b1t + s2e/2). In order to forecast the
number of Hispanic households for 2008 (t = 34), we compute
ŷ 34 = exp(7.9706 + 0.0479(34) + 0.03112/2 ) = 14,760.
Similarly, the forecast for Hispanic median income is computed as
ŷ34 = exp(9.2517 + 0.0418(34) + 0.07752/2 ) = $43,300.
Forecasts for other years can be computed similarly. Note that 2008 forecasts
with the exponential trend model are higher than those with the linear trend
model.
It is important to point out that whenever possible, it is preferred to use
unrounded values for making forecasts in an exponential model because even
a small difference, when exponentiated, can make a big difference in the
forecast.
c. We compute ŷ for the exponential model in regular units and not in logs. The
resulting ŷ also enables us to compare the linear and the exponential models in
terms of MSE and MAD or in terms of R2. In Table 18.11, we present a portion
of the seriesy„ along with y„ for both models; we did these calculations in
Excel with unrounded values for the estimates. We then compute
and where e, = y t – ŷt, While these calculations are not reported,
the MSE and MAD values for the linear and the exponential models are shown
in the last two rows of Table 18.11.
The exponential trend model appears to be better suited to describe the number
of Hispanic households, since it has a lower M SE and MAD than the linear
trend model. On the other hand, median income is better described by the lin
ear trend model. These findings are consistent with our earlier analysis with
Figures 18.4 and 18.5. Therefore, we use the exponential trend model to fore
cast the number of Hispanic households in 2008 as 14,760. The linear trend
model is used to forecast the median income of Hispanic households in 2008
as $37,980.
Polynomial Trends
Sometimes a time series reverses direction, due to any number of circumstances. A
common polynomial function that allows for curvature in the series is a quadratic
trend model. This model describes one change in the direction of a series and is esti
mated as
ŷt =β 0 + β1 t + β 2t2 + є t.
The coefficient β 2 determines whether the trend is U-shaped or inverted U-shaped.
Figure 18.6 depicts possible trends of a quadratic model.
In order to estimate the quadratic trend model, we generate t2, which is simply the
square of t. Then we run a multiple regression model that uses y as the response vari
able and both t and t2 as the explanatory variables. The estimated model is used to make
forecasts as
ŷ t = b0 + b 1t + b2t2.
FIGURE 18.6 Representative shapes o f a quadratic trend
In the cubic trend model we basically generate two additional variables, t2 and t3, for the
regression. A multiple regression model is run that uses y as the response variable and t,
t2, and t3 as the explanatory variables. The estimated model is used to make forecasts as
ŷ1 = b0 + b 1t + b2t2 + b3t3.
While we use the M SE and the MAD of in-sample forecast errors to compare the linear
and the exponential models, we cannot use them to compare the linear, quadratic, and
cubic trend models. The reason is that the values of MSE and MAD are always lowest for
the highest-order polynomial trend model, since the values decrease as we estimate ad
ditional parameters. The problem is similar to that of the coefficient of determination R2
discussed in earlier chapters. When comparing polynomial trend models, we use adjusted
R2, which imposes a penalty for over-parameterization.
THE P O L Y N O M IA L TREND M O DE L
A polynomial trend model of order q is estimated as
This model specializes to a linear trend model, quadratic trend model, and cubic
trend model for q = 1,2, and 3, respectively. The estimated model is used to make
forecasts a s ŷt = b0 + b 1t + b2t2 + b3t3 + • • • + bqtq, where b0, b 1, . . . , b q are the
coefficient estimates. We use adjusted R 2 to compare polynomial trend models with
different orders.
A good application of the polynomial trend model is used in the Writing with Statistics
section later in this chapter.
EXERCISES 18.3
Concepts a. Plot the series. Does the linear trend model or the
exponential trend m odel fit the series best?
10. Consider the fo llo w in g estim ated trend models. Use them to
make a forecast fo r t = 21. b. Estimate both models. Verify your answer in part a by
com paring the MSE o f the models.
a. LinearTrend:ŷ = 13.54 + 1.08t
c. Given the model o f best fit, make a forecast for visitors to
b. Q uadraticTrend: ŷ = 18.28 + 0.92t — 0.01 t2
am usement parks in 2008 and 2009.
c. Exponential Trend:
14. The p o te n tia lly deadly 2009 Swine Flu outbreak was
11 . F ILE Consider the fo llow ing table, consisting o f 20
due to a new flu strain o f subtype H1N1 no t previously
observations o f the variable y and tim e t . The data are also on
rep orted in pigs. W hen th e W orld Health O rganization
the text website, labeled Exercise 18.11.
declared a pandem ic, th e virus con tinu ed to spread in
t t t t
y y y y th e U nited States, causing illness along w ith regular
1 3.01 6 4.94 11 7.28 16 14.16
seasonal influenza viruses. Consider th e fo llo w in g 2009
2 3.13 7 6.10 12 9.04 17 14.85
w eekly data on to ta l Swine Flu cases in th e United
3 4.19 8 5.91 13 9.49 18 16.77
States, reported by the Centers fo r Disease C ontrol and
4 5.07 9 6.56 14 12.12 19 18.07
Prevention (CDC).
5 4.46 10 7.29 15 13.15 20 19.99
Week Total Week Total
a. Plot the series along w ith the superim posed linear and
17 1,190 22 2,904
exponential trends. W hich trend m odel do you th in k
describes the data well? 18 2,012 23 3,024
m odel fo r the sample. Validate your guess from the 20 2,247 25 1,829
graphs by com paring the MSE and MAD. 21 2,280 26 1,093
12 . F ILE Consider the follow ing table, consisting o f 20 SOURCE: w w w .c d c .g o v .
2010). Consider the follow ing table, which lists a portion a. Plot the above series. Which polynomial trend model do
of the m onthly unem ploym ent rates (seasonally adjusted) you th in k is most appropriate?
in California from 2007-2010. The full data set can be b. Verify your answer by form ally comparing the linear, the
found on the text website, labeled California
quadratic, and the cubic trend models.
U n em p lo y m e n t.
c. Forecast the unem ploym ent rate in California for January
2011.
Decomposition Analysis
Let T, S, and I represent the trend, the seasonal, and the random components, respec
tively, of the time series yt. With seasonal data, it is necessary to model seasonality along
with trend. Consider the following multiplicative model, which relates sample observa
tions to T, S, and I:
yt = Tt × St, × It.
We can use sample information to extract the trend and the seasonal components of the
series and then project them into the future. Let T, and S, denote the estimated trend and
seasonal components, respectively. Note that by their very nature, random variations can
not be identified. Therefore, we set Î, = 1 and make forecasts as ŷ=t Tt × Ŝt. This process
is often referred to as decomposition analysis. Alternatively, we can use a multiple re
gression model to simultaneously estimate trend along with seasonal dummy variables.
We first elaborate on decomposition analysis, then discuss the use of seasonal dummy
variables.
In the introductory case to this chapter, we considered the N ike Revenues data on FILE
Nike’s quarterly revenue from 1999 through 2008. Figure 18.8 is a scatterplot of the data,
where we have relabeled the 10 years of quarterly observations from 1 to 40. The graph
highlights some important characteristics of Nike’s revenue. First, there is a persistent
upward movement in the data. Second, the trend does not seem to be linear and is likely
to be better captured by an exponential model. Third, a seasonal pattern repeats itself year
after year. For instance, revenue is consistently higher in the first and fourth quarters as
compared to the second and third quarters.
We designate the first moving average ȳ2.5 because it represents the average in quarters 1
through 4. The next two moving averages, representing the average in quarters 2 through
5, and quarters 3 through 6 , respectively, are
The CMA series, tȳ,si shown in column 4 o f Table 18.12. (We should mention here that
Excel calculates ȳ2.5,ȳ3.5,ȳ4.5, . . . , when calculating a 4-quarter moving average. We then
need to prompt Excel to “ center” the values by creating another column that calculates
the average between each pairing.)
1999:02 2 1,913 — —
2008:04 40 5,088 — —
We note that ȳt eliminates seasonal variations and also some random variations, that is, ȳ,
represents a series that only includes the trend component Tt. Heuristically, since yt =
Tt, × S, × It and ȳt = Tt, when we divide y1,by ȳt, we are le ft w ith St, × I1. This series, y t/ ȳt,
is called the ratio-to-moving average and is presented in column 5 o f Table 18.12.
In Table 18.13, we rearrange y1, /ȳt, by quarter from 1999– 2008. Note that each quarter
has m ultiple ratios, where each ratio corresponds to a different year. For instance, yt/ȳt,
fo r the third quarter is 0.9924 in 1999, 0.9541 in 2000, and so on. In this example, each
quarter has nine ratios. We use the arithm etic average (sometimes statisticians prefer to
use the median) to determine a common value fo r each quarter. By averaging, we basi
cally cancel out the random component and extract the seasonal component o f the series.
We refer to this summary measure as the unadjusted seasonal index for the quarter. For
instance, the average o f the ratios fo r the first quarter is calculated as (1.1225 + • • • +
1.0571)/9 = 1.0815.
For quarterly data, the seasonal indices must add up to 4 (the number o f seasons or m) so
that the average is one. In order to ensure this requirement, we m ultiply each unadjusted
seasonal index by 4 and divide by the sum o f the four unadjusted seasonal indices. In this
case, the “ m ultiplier” equals
Therefore, the adjusted seasonal index fo r quarter 1 is calculated as 1.0815(1.0004) =
1.0819. This is our final estimate o f the seasonal index rounded to the 4th decimal
place— it is referred to as the adjusted seasonal index. Table 18.13 shows adjusted
seasonal indices fo r each quarter. Note that the average o f the adjusted seasonal indices
equals one.
Let us interpret these adjusted seasonal indices. There is no seasonality i f all in d i
ces equal their average o f one. On the other hand, i f the seasonal index fo r a quarter is
greater (less) than one, it im plies that the observations in the given quarter are greater
(less) than the average quarterly values. In the above example, the adjusted seasonal
index o f 1.0819 fo r quarter 1 im plies that N ike’s revenue is about 8.19% higher in
the first quarter as compared to the average quarterly revenue. The adjusted seasonal
index fo r the second quarter is 0.9417, suggesting that revenues are about 5.83% lower
than the average quarterly revenue. Other adjusted seasonal indices are interpreted
sim ilarly.
As noted earlier, the adjusted seasonal index o f 1.0819 implies that N ike’s revenue in the
first quarter is about 8.19% higher than average quarterly revenue. Therefore, w ithout the
seasonal effect, the revenue would be lower. For instance, the revenue o f 2,505 m illion
in 1999:01 is only 2,315.29 m illion once it has been seasonally adjusted (we used un
rounded values in the calculations). In Figure 18.9, we plot revenue along w ith seasonally
adjusted revenue.
Note that the seasonally adjusted series is free o f seasonal variations, thus highlight
ing the long-term movement (trend) o f the series. Figure 18.9 also confirms that the expo
nential trend model is better suited fo r the data than is the linear trend model.
FIGURE 18.9 Regular and seasonally adjusted
quarterly revenue
In order to obtain the seasonally adjusted series, we divide the original series by
its corresponding seasonal index (y ,/Ŝ t). We use the appropriate trend model on the
seasonally adjusted series to extract the trend component
We estimate the exponential trend model where the response variable is the natural log of
the seasonally adjusted revenues ln(yt/ Ŝt,) and the explanatory variable t assumes values
1 , 2 , . . . , 40, representing 10 years of quarterly data from 1999 to 2008. The relevant
portion of the regression output is given in Table 18.15. We encourage you to compare the
results of this model with the linear trend model, using model selection criteria.
TABLE 18.15 Exponential Regression Results on Seasonally Adjusted Data
Coefficients Standard Error t Stat p-value
Note that the slope coefficient is positive and highly significant (p-value ≈ 0). We
forecast trend for the seasonally adjusted revenue as There
fore, a trend forecast for the seasonally adjusted revenue in the first quarter of 2009 is
computed as
EXAMPLE 18.5
A tourism specialist uses decomposition analysis to examine hotel occupancy rates
for Bar Harbor, Maine. She collects quarterly data for the past five years (n = 20)
and finds that the linear trend model best captures the trend of the seasonally
adjusted series: In addition, she calculates quarterly indices as
Ŝ = 0.53, Ŝ2 = 0.90, Ŝ 3 = 1.40, and Ŝ 4 = 1.17.
a. Interpret the first and third quarterly indices.
b. Forecast next year’s occupancy rates.
SOLUTION:
a. The first quarter’s index of 0.53 implies that the occupancy rate in the first quar
ter is 47% below the average quarterly occupancy rate, whereas the occupancy
rate in the third quarter is 40% above the average quarterly occupancy rate.
b. We calculate the next year’s quarterly occupancy rates as
Year 6, Quarter 1 : ŷ2l = (0.60 + 0.0049(21)) × 0.53 = 0.3725.
Year 6, Quarter 2: ŷ22 = (0.60 + 0.0049(22)) × 0.90 = 0.6370.
Year 6, Quarter 3: ŷ23 = (0.60 + 0.0049(23)) × 1.40 = 0.9978.
Year 6, Quarter 4: ŷ 24 = (0.60 + 0.0049(24)) × 1.17 = 0.8396.
The estimated equation, w ith values rounded to the 4th decimal place, is ŷ =
exp(7.5929 + 0.050 d 1,– 0.1036d2 –– 0.0985d3 + 0.0218t + 0.05972/2 ). The coef
ficients for seasonal dummy variables indicate that the revenue is about 5% higher
in the first quarter and about 10% lower in the second and third quarters as com
pared to the fourth quarter. The trend coefficient suggests that the quarterly increase
in revenue is about 2.18%.
For 2009:01, we use d 1 = 1, d2 = 0, d3 = 0, t = 41 to forecast N ike’s revenue
as ŷ41 = exp(7.5929 + 0.0501 + 0.0218(41) + 0.05972/2 ) = $5,108.10 million.
Similarly, for 2009:02, we use d 1 = 0, d2 = 1, d3 = 0, t = 42 to determine ŷ42 =
exp(7.5929 – 0.1036 + 0.0218(42) + 0 .0 5 9 7 7 2 ) = $4,476.88 million.
Forecasts for 2009:03 and 2009:04 yield $4,598.94 million and $5,186.85 million,
respectively. As before, w henever possible, it is advisable to make forecasts with
unrounded values.
As em phasized earlier, we always use model selection criteria to choose the ap
propriate forecasting model. In Table 18.18 we present the M S E and M A D based on
the residuals, et = yt — ŷ t ,with decom position analysis and seasonal dum m y vari
ables; we did these calculations in Excel w ith unrounded values for the estimates.
We encourage you to replicate these results.
We select the decomposition analysis method to make forecasts because it has the
lower MSE and MAD of in-sample forecast errors. Therefore, the quarterly forecasts
for 2009, as derived earlier, are $5,070.89, $4,511.04, $4,593.82, and $5,194.44
million, respectively. This results in a sum of $19,370.19 million or $19.37 billion
for fiscal year 2009.
SYNOPSIS OF I N T R O D U C T O R Y CASE
17. Six years o f quarterly data o f a seasonally adjusted series are 1 6.49 7.34 7.11 10.82
19. Ten years o f m onthly data o f a seasonally adjusted series are d. Use the appropriate model to make forecasts for all four
used to estimate a linear trend model as quarters o f the next year.
In addition, seasonal indices for January and February are 23. FILE Consider a portion o f m onthly sales data for 5 years
calculated as 1.04 and 0.92, respectively. Make a forecast for for a growing firm . The full data set can be found on the text
the first tw o m onths o f next year. website, labeled Exercise 18.23.
20. FILE Consider the follow ing 20 observations, representing Year M onth Sales
quarterly inform ation for 5 years. The data are also on the 1 Jan 345
text website, labeled Exercise 18.20.
1 Feb 322
Year Q uarter 1 Q uarter 2 Q uarter 3 Q uarter 4 ⋮ ⋮ ⋮
1 8.37 12.78 8.84 15.93 5 Dec 10,745
2 10.03 12.48 8.91 24.81
a. Construct the seasonal indices for the data.
3 9.61 9.65 15.93 22.00
b. Plot the seasonally adjusted series to recommend the
4 8.80 11.45 6.79 10.16
appropriate trend model.
5 7.46 10.58 13.35 19.77
c. Use the trend and seasonal estimates to make forecasts
a. Calculate the 4-quarter centered m oving average. for the next tw o months.
b. Calculate the ratio-to-m oving average. 24. FILE Revisit the Exercise 18.23 data to estimate (a) a linear
c. Calculate and interpret the seasonal indices for quarters 1 trend model w ith seasonal dummies, (b) an exponential trend
and 4. model w ith seasonal dummies. Which o f the tw o models has
a lower MSE and M AD ? Use the appropriate model to make
21. FILE Consider the follow ing m onthly observations for
forecasts for the next tw o months.
5 years. The data are also on the te xt website, labeled
Exercise 18.21.
Applications
Year Jan Feb M ar Apr M ay Jun Jul Aug Sep Oct Nov Dec
25. FILE Hybrid cars have gained popularity because o f their fuel
1 13 17 15 12 32 15 21 17 33 15 34 21 economy and the uncertainty regarding the price o f gasoline.
2 26 10 14 19 17 14 27 30 25 15 18 27 All automakers, including the Ford M otor Co., have planned
3 24 11 19 23 19 18 31 18 15 31 34 27 to significantly expand their hybrid vehicle lineup (CNN.com,
4 23 22 17 24 17 24 15 19 33 19 16 39 November 9,2005).The follow ing table contains quarterly
5 17 19 25 30 18 17 19 36 19 26 21 35 sales o f Ford and Mercury hybrid cars. The data are also on the
text website, labeled Sales Data.
a. Calculate the 12-m onth centered m oving average.
Year Q1 Q2 Q3 Q4
b. Calculate the ratio-to-m oving average.
2006 6,192 5,663 4,626 5,645
c. Calculate and interpret the seasonal indices for April and
2007 5,149 6,272 5,196 6,545
November.
2008 5,467 5,235 3,160 7,007
22. FILE Consider the follow ing 20 observations, representing
quarterly inform ation for 5 years.The data are also on the 2009 5,337
text website, labeled Exercise 18.22. SOURCE: Internal Revenue Service, United States Department o f Treasury.
a. Plot the above series. Comment on the trend and formats of movies at Blockbuster Inc. (in millions o f dollars).
seasonal variation in the sales o f hybrid cars. The full data set can be found on the text website, labeled
b. Compute and interpret the seasonal indices. Blockbuster.
a. Estimate a linear trend model w ith seasonal dum m y SOURCE: Federal Reserve Bank of St. Louis.
yt = ß 0 + ß1xt + єt.
H ere y is the response variable caused by the explanatory variable x. L et the sam ple
observations be denoted by y 1, y2, . . ., y Tand x 1, x 2, . . . , x T, respectively. O nce w e have
estim ated this m odel, w e can m ake a one-step-ahead forecast as
ŷT+1 = b 0 + b 1 xT+1.
M ulti-step-ahead forecasts can be m ade sim ilarly. T his causal approach w orks only if we
know, or can predict, the future value o f the explanatory variable x T+1. For instance, let
sales y be related to expenditure on advertisem ent x . We can forecast sales ŷT+l only if we
know the advertisem ent budget, x T+1 fo r the next period.
y t = ß 0 + ß1 xt-t + є t.
w here ß1, represents the slope o f the lagged explanatory variable x. N ote that if w e have
T sam ple observations, the estim able sam ple w ill consist o f T — 1 observations, w here
y2, y 3, . . . , y T are m atched w ith x 1, x 2, . . . , x T-1. H ere a one-step-ahead forecast is easily
m ade as
ŷT+1 = b0+b1 xT .
T his fo recast is not conditional on any predicted value o f the explanatory variables,
since x T is its last know n sam ple value. We can g eneralize this m odel to include m ore
lags o f x. For exam ple, w e can specify a tw o-period lagged regression m odel as yt =
ß 0 + ß1xt-1 + ß2xt- 2 + єt. A o ne-step-ahead forecast is now m ade as ŷ T+1 = b0 +
b 1xT + b2x T- 1.
A nother popular specification for causal forecasting uses lagged values o f the re
sponse variable as an explanatory variable. For instance, consider the m odel
y t = ß0 + ß1yt-1 + є t.
w here the param eter ß t represents the slope o f the lagged response variable y. T his re
gression is also referred to as an autoregressive model o f order one, or sim ply an AR( 1).
H igher-order autoregressive m odels can be constructed sim ilarly. A one-period-ahead
forecast is m ade as
ŷT+1 = b0 + b1yT.
Finally, w e can also use lagged values o f both x and y as the explanatory variables. For
instance, consider
Note that we lose one observation fo r the regression, since we do not have inform a
tion on the lagged values fo r 1981. Table 18.21 summarizes the regression results
o f the three models.
TABLE 18.21 M odel Evaluation w ith Causal Forecasting
Therefore, we forecast that about 412,490 net private housing units w ill be sold in 2009.
EXERCISES 18.5
Concepts a. Estimate y1
, = ß0 + ß
t-
X
1 + єt.
33. FILE Consider the follow ing portion o f data on the response
b. Estimateyt = ß0 + ß1yt-1 + є t
va riabley and the explanatory variable x. The data are on the c. Estimate yt = ß0 + ß1Xt-1 + ß2yt-1 + єt
te xt website, labeled Exercise 18.33. d. Use the most suitable model to make a one-step-ahead
t y x forecast (t = 13) for y.
1 27.96 29.88
2 30.15 21.07 Applications
37. Hiroshi Sato, an ow ner o f a sushi restaurant in San
12 24.47 26.41 Francisco, has been fo llo w in g an aggressive m arketing
a. Estimate y ,t = ß 0 + ß 1 X t - 1 t+
є
and y t = ß0 + ß1Xt-1 + cam paign to th w a rt th e effect o f rising unem ploym ent
rates on business. He used m o n th ly data on sales
ß2Xt-2 + єt.
($1,000s), advertising costs ($), and th e unem ploym ent
b. Use th e appropriate m odel to make a one-step-ahead
rate (%) from January 2008 to May 2009 to estim ate the
forecast (t = 13) fo r y.
fo llo w in g sam ple regression equation:
34. F IL E Consider the follow ing po rtion o f data on the
variable y. The data are on the te x t website, labeled
Exercise 18.34.
t y a. Hiroshi had budgeted $620 tow ard advertising
costs in May 2009. Make a forecast for Sales for
1 29.32
June 2009 if the unem ploym ent rate in May 2009
2 30.96
was 9.1%.
⋮ ⋮
b. W hat w ill be the forecast if he raises his advertisem ent
24 48.58
budget to $700?
a. Estimate an autoregressive model o f order 1, yt = ß 0 + c. Reevaluate the above forecasts if the unem ploym ent
ß1yt-1 + є t , to make a one-step-ahead forecast (t = 25) fo r y. rate was 9.5% in May 2009.
b. Estimate an autoregressive m odel o f order 2, y t = β 0 + 38. FILE The P hillips curve is regarded as a reliable
β1 y t_1 + β 2 y t - 2 + єt, to make a one-step-ahead forecast to o l fo r forecasting in fla tio n . It captures th e inverse
(t = 25) fo r y. re lation betw een th e rate o f u n em p loym e nt and the
c. W hich o f the above models is more appropriate for rate o f in fla tio n ; th e lo w er the u n em p loym e nt in an
forecasts? Explain. econom y, th e high er is th e in fla tio n rate. Consider
35. F IL E Consider the follow ing portion o f data on y and x that th e fo llo w in g p o rtio n o f m o n th ly data on seasonally
appears on the text website, labeled Exercise 18.35. adjusted in fla tio n and u n em p loym e nt rates in the
c. Which of the above models is more appropriate for a. Estimate tw o models, o f order 1 and 2, using
forecasts? Explain. unem ploym ent as the response variable and lagged
36. E
IL
FConsider the fo llow ing po rtion o f data on y and x tha t inflation as the explanatory variable(s). Should you
appears on the te xt website, labeled Exercise 18.36. use either model for forecasting unem ploym ent?
t y x Explain.
1. Evaluate various polynom ial trend models for the inflation rate.
2. Use the best-fitting trend model to forecast the inflation rate for June 2009.
Economists generally agree that high levels of inflation are caused by the money sup
Sample ply growing faster than the rate of economic growth. During high inflationary pressures,
monetary authorities decrease the money supply, thereby raising short-term interest rates.
Report— Sometimes they also have to contend with deflation, or a prolonged reduction in the level
of prices. As prices fall, consumers tend to delay purchases until prices fall further, which
Forecasting in turn can depress overall economic activity.
The global economic crisis that began in the summer/fall of 2008 raised defla
the Monthly tionary fears, with rapidly rising unemployment rates and capital markets in turmoil.
An increase in price levels in 2007 was followed by a decrease in 2008 with a slight
Inflation Rate hint of price stability in the second quarter o f 2009. This report does not focus on the
effectiveness of monetary policy. Instead, a forecast of the inflation rate is made from
a noncausal perspective by simply projecting historical data. Seasonality is not a con
cern, since the inflation data are already seasonally adjusted.
A simple plot of the inflation rate from January 2007 to May 2009 is shown in
Figure 18.A. In order to gauge whether a linear or nonlinear trend is appropriate, vari
ous trend models are superimposed on the inflation rate scatterplot. The exponential
trend is not included, since the log of the inflation rate is not defined for nonpositive
inflation rates.
Interestingly, the implied forecasts seem to vary widely between competing models.
Although Figure 18.A suggests that the cubic model accurately captures the changing
trend of the inflation rate over the last 29 months, this finding must be supplemented with
formal model selection criteria.
Three trend models were estimated, where yt represents the inflation rate and t rep
resents the relabeled monthly observations from 1 (January 2007) through 29 (May
2009).
Linear Model: y t =ß 0 + ß1t + єt
Quadratic Model: yt = ß0 + ß1t + ß 2t2 + єt
Cubic Model: yt = ß0 + ß1t + ß 2t2 + ß 3t3 + єt
Table 18.A presents parameter estimates of the three models. Also included in
the table are the adjusted R2 for model comparison, which suitably penalizes over
parameterization.
TABLE 18.A Analysis o f the Linear, Quadratic, and Cubic Trend Models
Variable Linear Quadratic Cubic
Constant 0.5153* NA NA
(0.011)
t – 0.0229* 0.0583* 0.1472*
(0.047) (0.012) (0.009)
t2 NA – 0.0025* – 0.0125*
(0.014) (0.029)
t3 NA NA 0.0003*
(0.072)
Adjusted R2 0.1058 0.1463 0.2143
N otes: The top portion o f the table contains parameter estimates w ith p-values in parentheses; N A denotes
not applicable; the constant is removed from the quadratic and cubic trend models since it is found to
be insignificant; the symbol * represents significance at the 10% level; the last row o f the table contains
adjusted R2 for model comparison.
Consistent w ith the inform al graphical analysis, the cubic trend provides the best sample
fit, as it has the highest adjusted R2 o f 0.2143. Therefore, the estimated cubic trend model
is used w ith unrounded estimates to derive the forecast fo r June 2009 as
Inflation forecasts are w idely regarded as key inputs fo r implementing monetary policy.
Whether or not it can be forecasted accurately is an open empirical question. In this re
port, an attempt is made to forecast inflation, despite market turbulence.
Conceptual Review
LO 18.1 Distinguish among the various models used in forecasting.
Observations o f any variable recorded over time in sequential order are considered a
tim e series. The purpose o f any forecasting model is to forecast the outcome o f a time
series at time t, or ŷ t . Forecasting methods are broadly classified as quantita tive or
qualitative. W hile qualitative forecasts are based on prior experience and the expertise
o f the forecaster, quantitative forecasts use a form al model, along w ith historical data
fo r the variable o f interest. Quantitative forecasting models are further divided into
causal and noncausal models. Causal methods are based on a regression framework,
where the variable o f interest is related to a single or m ultiple explanatory variables.
Noncausal models, also referred to as purely time series models, do not present any
explanation o f the mechanism generating the variable o f interest and sim ply provide a
method fo r projecting historical data.
The in-sample forecast y, is also called the predicted or fitte d value o f y t The in-sample
forecast errors or the residuals are computed as et = yt — ŷ t . We use the residuals,
computed as e, = y, — y„ to compute When selecting
among various models, we choose the one w ith the smallest MSE and MAD. I f MSE
and MAD provide conflicting results, then we choose the model w ith the smallest MSE.
LO 18.4 Calculate and interpret seasonal indices and use them to seasonally adjust a
time series.
Centered Moving Averages CMA are often employed to separate the effect of trend
from seasonality. We use a centered 4-period moving average for quarterly data and a
centered 12-period moving average for monthly data. The ratio-to-moving average is
calculated as y ,t / ȳ t where ȳ t = CMAt. For each season, there are many estimates of the
ratio-to-moving average. The unadjusted seasonal index is computed by averaging
these ratios over the years. A minor adjustment ensures that the average of adjusted
seasonal indices Ŝ equals one. A seasonally adjusted series is computed as yt/ Ŝ t . We
use the appropriate trend model on the seasonally adjusted series to extract
LO 18.6 Use trend regression models with seasonal dummy variables to make forecasts.
As an alternative to decomposition analysis, we use a multiple regression model to
simultaneously estimate trend along with seasonal dummy variables. With quarterly
data, a linear trend model with seasonal dummy variables is specified as y = + ß0ß1d1
+ ß 2d2 + ß3d3 + ß4t + є . Forecasts based on the estimated model are ŷt = (b0 + b 1) +
b4t (Quarter 1), ŷt = (b0 + b2) + b4t (Quarter 2), ŷt = (b0 + b3) + b4t (Quarter 3), and
ŷ t = b0 + b4t (Quarter 4). An exponential trend model with seasonal dummy variables
is specified as ln(y) = β 0 + β 1 d 1 + β 2d2 + β 3d3 + β4 t + є . Forecasts based on the
estimated model are ŷt = exp((b0 + b 1) + b4t + s2e/2 ) (Quarter 1), ŷt = exp((b0 + b2) +
b4t + s2 e ) (Quarter 2), ŷ = exp((b0 + b3) + b4t + s2e/2 ) (Quarter 3), and ŷt = exp (b0 +
b4t + s2e/ 2) (Quarter 4). Forecasts with monthly data can be made similarly.
A d d itio n a l E x e r c is e s a n d C a s e S tu d ie s
40. FILE The U.S. housing market remains fragile despite a. Plot each of the above series and comment on the
historically low mortgage rates (AARP, July 2, 2010). Since respective trend.
the rate on 30-year mortgages is tied to the 10-year yield on b. Estimate a linear, a quadratic, and a cubic trend
Treasury bonds, it is important to be able to predict this yield model for the share of country music in the
accurately. The accompanying table shows a portion of the United States. Use their adjusted R2to choose
10-year yield on Treasury bonds (in %) for 21 trading days in the preferred model and with this model make a
November 2010; the complete data set can be found on the forecast for 2009.
text website, labeled Yields. c. Estimate a linear, a quadratic, and a cubic trend
Date Yield (in %) model for the share of rap/hip-hop music in the
1-Nov 2.63 United States. Use their adjusted R2to choose the
2-Nov 2.59 preferred model and with this model make a forecast
⋮ ⋮ for 2009.
30-Nov 2.80 42. FILE Tourism was hit hard by the international financial
crisis that began in the fall of 2008. According to the
SOURCE: f i n a n c e . y a h o o . c o m .
Bureau of Economic Analysis (December 20, 2010), tourism
a. Use a 3-period moving average to make a forecast for spending has picked up, but it still remains below its peak,
December 1, 2010. which occurred in 2007.The accompanying table shows
b. Use the exponential smoothing method to make a a portion of seasonally adjusted data on real tourism
forecast for December 1, 2010. Use α = 0.5. spending (in millions of $); the full data set is on the text
c. Which of these smoothing methods has a better in- website, labeled Tourism Spending.
sample performance?
Year Quarter 1 Quarter 2 Quarter 3 Quarter 4
d. The actual 10-year yield on December 1, 2010 was 2.96.
Was the forecast performance of the two methods 2004 664,924 672,678 675,262 679,410
consistent with their in-sample performance in part c? 2005 683,989 690,226 695,753 700,037
41. FILE The following table lists a portion of the percentage (share) ⋮ ⋮ ⋮ ⋮ ⋮
of total shipment of music that falls in the category of country 2010 676,929 682,681 695,917 NA
and rap/hip-hop rock music from 1990–2008. The complete
SOURCE:U.S. Bureau of Economic Analysis.
data set is on the text website, labeled Country and Rap.
Year Country (in %) Rap/Hip-hop (in %) a. Construct seasonal indices for the data.
1990 9.6 8.5 b. Plot the seasonally adjusted series. Estimate the cubic
trend model on the seasonally adjusted series.
1991 12.8 10.0
⋮ ⋮ ⋮ c. Use the seasonal and trend components to
forecast tourism spending for the fourth quarter of
2008 11.9 10.7 2010 and the first three quarters of 2011.
SOURCE: www.riaa.com.
43. FILE Prices of crude oil have been steadily rising over fu ll data set can be fou nd on the te x t website, labeled
the last tw o years (The Wall Street Journal, December 14, In v e n to ry -to -S a le s.
b. Compare the adjusted R2 of the linear, the quadratic, and m u ltip ly in g the num ber o f paying passengers by the
the cubic trend models. distance flo w n in thousands. The accom panying table
shows a p o rtio n o f m o n th ly data on revenue passenger-
c. Use the appropriate model to make a forecast for the
miles (in m illions) from January 2006 th ro u g h Septem ber
price o f regular unleaded gasoline for January and
2010; the fu ll data set can be fou nd on th e te x t website,
February o f 2011.
labeled Revenue Passenger-Miles.
44. FILE Consider the following portion of data on real estate
loans granted by FDIC-insured Commercial Banks in the Year M onth Revenue Passenger-Miles
United States (in billions of U.S. dollars, base = 2007) from 2006 Jan 43.1652
1972 to 2007; the full data set can be found on the text Feb 44.0447
2006
website, labeled Loans. ⋮ ⋮ ⋮
2010 Sep 43.7704
Year Loans
1973 567.26 a. Plot the above series and com m ent on its trend and
⋮ ⋮ seasonal variations.
year 2008 w ith this model. c. Use the MSE and M A D to compare these models.
d. Use the appropriate model to make m onthly forecasts
45. FILE W hile U.S. in ven to ry levels rem ain low, there is
for the last three months of 2010.
a slight in dica tion o f an increase in the U.S. business
inventory-to-sales ratio, due to higher sales (The W a ll 48. F IL E follow ing data represent a po rtion o f quarterly
S tre e t J o u rn a l, December 15, 2010). The accom panying net sales (in m illions o f dollars) o f Lowe's Companies, Inc.,
table shows a p o rtio n o f seasonally adjusted in ven to ry- over the past five years; the full data set can be found on
to-sales ratios from January 2008 to October 2010; the the te x t website, labeled Lowe's N e t Sales.
Year Q uarter 1 Q uarter 2 Q uarter 3 Q uarter 4 Estimate three models: (a) yt = β 0 + β1xt − 1 + єt, (b) yt =
firm is atte m p tin g to forecast the daily stock price o f 2006 1 9705.2 9148.2
Genzyme Corporation, one o f the world's leading biotech 2006 2 9863.8 9266.6
companies, using both the S&P 500 Index as well as ⋮ ⋮ ⋮ ⋮
Genzyme's past stock prices. The follo w in g table shows a
2010 4 11514.7 10525.2
portion o f the daily adjusted closing prices o f Genzyme (y )
SOURCE: U.S. D e pa rtm e n t o f Commerce.
and the S&P 500 Index (x ) from December 1,2010, to
December 22, 2010; the full data set is on the te xt website, a. Plot the consumption series. Estimate the appropriate
labeled G e n zym e . polynomial trend model to forecast consumption
expenditure for the 1st quarter o f 2011.
Date y X
b. Estimate yt = β 0 + β1xt - 1+ ε t to forecast consumption
12/ 1/2010 71.14 1206.07 expenditure for the 1st quarter o f 2011.
12/ 2/2010 70.97 1221.53 c. Which o f these tw o models is more appropriate for
⋮ ⋮ ⋮ making forecasts? Explain.
12/22/2010 71.52 1258.84
SOURCE: www.finance.yahoo.
CASE STUDIES
1. Construct the exponentially smoothed series for fried dough and soft drinks using
α = 0.30 and α = 0.70.
2. Calculate MSE and MAD fo r each series.
3. Forecast sales o f fried dough and soft drinks fo r day 21 w ith the best-fitting series.
1. Use a scatterplot to determine which model best depicts trend fo r W al-M art’s sales.
2. Determine whether or not a seasonal component exists in the series, using the sea
sonal dummy variable approach.
3. Given the conclusions on trend and the seasonal component, provide forecast values
fo r the four quarters o f 2011 as w ell as total projected sales fo r fiscal year 2011.
LEARNING OBJECTIVES
E R
R e tu rn s ,
Index
Numbers, and Inflation
After readingLO 19.1 Define and compute investment returns.
this chapter
you shouldLO 19.2 Use the Fisher equation to convert nominal returns into real returns and
be able to: vice versa.
LO 19.5 Compare the Laspeyres and the Paasche methods for computing the
weighted aggregate price index.
LO 19.6 Use price indices to deflate economic time series and derive the inflation rate.
In Chapter 18 we derived seasonal indices to adjust tim e series for seasonal changes. Policy makers
often analyze tim e series in this format, as they are not particularly interested in its seasonal varia
tions. Other transformations o f tim e series also facilitate interpretation and statistical analysis. For
example, financial analysts are interested in the analysis o f investm ent returns. The underlying data
may consist o f asset prices and income distributions, but these can easily be transformed into invest
m ent returns. Similarly, economists are often interested in measuring the m agnitude o f economic
changes over time. They can create index numbers tha t transform the original data into figures rep
resenting percentage changes. Finally, many tim e series are reported both in nominal as well as real
terms. While the nominal values represent dollar amounts, the corresponding real values incorporate
inflation to represent the purchasing power o f money. In this chapter we will com pute and interpret
where Pt and Pt−1 are the price o f the asset at times t (current) and t − 1 (prior),
respectively, and It is the income distributed during the investment period. The
ratios are the capital gains yield and the income yield compo
nents, respectively.
The process fo r computing an investment return is the same fo r all assets. The income
component is dividends fo r stocks, interest fo r bonds, and rental income fo r a real estate
investment. For some assets, like Treasury bills, there is no income component and the
investment return consists entirely o f a capital gain or loss.
EXAMPLE 19.1
Helen Watson purchased a corporate bond fo r $950 a year ago. She received a
coupon payment (interest) o f $60 during the year. The bond is currently selling for
$975. Compute Helen’s (a) capital gains yield, (b) income yield, and (c) investment
return.
SOLUTION:
c. The investment return is the sum o f the capital gains yield and the income
yield, that is, 0.0263 + 0.0632 = 0.0895 or 8.95%. We can also compute it
directly as
EXAMPLE 19.2
Last year Jim Hamilton bought a stock for $35 and recently received a dividend
of $1.25. The stock is now selling for $31. Find Jim ’s (a) capital gains yield,
(b) income yield, and (c) investment return.
SOLUTION:
a. The capital gains yield is
Note that the investment return is unaffected by the decision to sell or hold assets. A
common misconception is that if you do not sell an asset, there is no capital gain or loss
involved, as a given price increase or decrease leads only to paper gain or loss. This mis
conception often leads an investor to hold a “loser” asset longer than necessary because
of the reluctance to admit a bad investment decision. The nonrecognition of the loss is
relevant for tax purposes, since only realized income must be reported in tax returns.
However, whether or not you have liquidated an asset is irrelevant when measuring its
performance.
U S I N G A D J U S T E D C L O S E P RI CE S T O
CALCULATE AN IN V E S T M E N T RETURN
Let P*t and P*t−1 represent the adjusted close price of a stock at times t (current)
and t − 1 (prior), respectively. The investment return R, at the end of time t is cal
culated as
EXAMPLE 19.3
Consider the adjusted close stock prices o f M icrosoft Corporation in Table 19.2.
Find the m onthly returns fo r November and December o f 2010.
TABLE 19.2 M on th ly Stock Prices for M icrosoft Corporation
THE FISHER E Q U A T IO N
Let R be the nominal rate o f return, r the real rate o f return, and i the expected infla
tion rate. The Fisher equation is defined as
When the expected inflation rate is relatively low, a reasonable approximation to the
Fisher equation is r = R − i; we w ill not be using this approximation in this chapter.
EXAMPLE 19.4
The quoted rate o f return on a one-year U.S. Treasury b ill in January 2010 is 0.45%
(www.ustreas.gov). Compute and interpret the real rate o f return that investors can
earn i f the inflation rate is expected to be 1.6%.
EXERCISES 1 9. 1
1. You borrowed $2,000 to take a vacation in the Caribbean 7. You w ould like to invest $20,000 fo r a year in a risk-free
islands. At the end o f the year, you had to pay back $2,200. investm ent. A conventional CD offers a 4.6% annual rate
What is the annual interest tha t you paid on your loan? o f return. You are also considering an "Inflation-Plus"CD
2. You bought a corporate bond last year for $980. You received which offers a real rate o f return o f 2.2% regardless o f the
a coupon payment (interest) o f $60 and the bond is currently inflation rate.
selling for $990. What is the (a) income yield, (b) capital gains a. What is the implied (expected) inflation rate?
yield, and (c) total return on the investment? b. You decide to invest $10,000 in the conventional and
3. The price o f a stock has gone up from $24 to $35 in $10,000 in the "Inflation-Plus" CD. What is your expected
one year. It also paid a year-end dividend o f $1.20. What dollar value at the end o f the year?
is the stock's (a) income yield, (b) capital gains yield, and c. Which o f the tw o CDs is a better investm ent if the
(c) total return? actual inflation rate for the year turns out to be
2 .2 %?
4. The year-end price and dividend inform ation on a stock is
given in the follow ing table. 8. Consider the follow ing adjusted close stock prices o f Intel
Year Price Dividend Corporation. Find the m onthly returns for November and
December o f 2010.
1 $23.50 NA
2 24.80 $0.18 Date Adjusted Close Price
3 22.90 0.12 December 2010 $21.48
a. What is the nominal return ot the stock in years 2 November 2010 $20.98
b. What is the corresponding real return if the inflation rates SOURCE: Data obtained from http://finance.yahoo.com on December 2010.
for years 2 and 3 were 2.8% and 1.6%, respectively? 9. Consider the follow ing adjusted close stock prices
5. A portfolio manager invested $1,500,000 in bonds in 2007. of Johnson and Johnson (J&J) and Caterpillar, Inc.
In one year the market value o f the bonds dropped to Compute and compare the m onthly returns for both
$1,485,000. The interest payments during the year totaled companies.
$105,000.
Date Johnson and Johnson Caterpillar
a. What was the manager's total rate o f return for the year?
March 2011 60.70 99.86
b. What was the manager's real rate o f return if the inflation
February 2011 61.44 102.93
rate during the year was 2.3%?
January 2011 59.23 97.01
6. Bill Anderson purchased 1,000 shares of Microsoft Corporation
December 2010 61.30 93.22
stock for $17,100 at the beginning o f 2009. At the end of the
year, he sold all o f his Microsoft shares at $30.48 a share. He November 2010 61.00 84.20
also earned a dividend o f $0.52 per share during the year. October 2010 62.63 78.23
a. What is Bill's total return on the investment? SOURCE: Data obtained from http://finance.yahoo.com on March 2011.
A S I M P L E P RI CE I N D E X
A simple price index for any item is the ratio of the price in period t, p t, and the
price in the base period. p Q, expressed as a percentage. It is calculated
EXAMPLE 19.6
Consider the data presented in the introductory case of this chapter in Table 19.1. Use the
base year of 2007 to compute and interpret the 2008 and 2009 simple price indices for:
a. red wine
b. white wine
c. 6-pack of beer
SOLUTION: Since 2007 is the base year, we set the corresponding index value equal
to 100. The index values for other years are computed below,
a. For red wine, the simple price index for 2008 is
Therefore, the average price of red wine in 2008 and 2009 was 98.37% and
80.89%, respectively, of what it was in 2007. In other words, as compared to
2007, the price of red wine dropped by 1.63% in 2008 and 19.11% in 2009.
b. For white wine, the simple price index for 2008 is (11.05/11.90) X 100 = 92.86
for 2008 and (10.60/11.90) X 100 = 89.08 for 2009. Therefore, relative to 2007,
the average price of white wine dropped by 7.14% in 2008 and 10.92% in 2009.
c. The simple price index for a six-pack of beer is (8.25/8.10) X 100 = 101.85
for 2008 and (7.95/8.10) X 100 = 98.15 for 2009. Interestingly, while the
prices of both red and white wines experienced substantial declines, the price
of beer stayed fairly stable. Relative to the base year of 2007, there was a
1.85% increase in the price of beer in 2008 and a 1.85% decline in 2009.
EXAMPLE 19.7
Table 19.3 shows the average price and corresponding price index fo r gasoline from
2000 to 2008. Interpret the price indices fo r 2001 and 2008.
TABLE 19.3 Price and Corresponding Price Index for Unleaded Gasoline in U.S., Base
Year 2000
Year 2000 2001 2002 2003 2004 2005 2006 2007 2008
Price 1.51 1.46 1.36 1.59 1.88 2.30 2.59 2.80 3.27
Price Index
(Base = 2000) 100 96.69 90.07 105.30 124.50 152.32 171.52 185.43 216.56
SOLUTION: Since 2000 is treated as the base year, the index number fo r 2000 is
100. The index number fo r 2001 is calculated as (1.46/1.51) X 100 = 96.69. Thus,
the gasoline price in 2001 was 96.69% o f what it was in 2000, or 3.31% lower.
Given a price index o f 216.56 in 2008, the gasoline price in 2008 was 116.56%
higher relative to 2000.
In Figure 19.1, we plot the raw price and price indices for gasoline from 2000 to 2008.
Note that although the units o f the gasoline price and index number graphs are different,
the basic shape o f the two graphs is similar. This shows that the main purpose o f index
numbers is to provide an easy interpretation o f the changes o f the series over time.
FIGURE 19.1 Price o f gasoline and the corresponding index numbers for 2000– 2008
It is im portant to note that index numbers provide direct comparisons only w ith respect
to the base year. S im ilar direct comparisons cannot be made between non-base years. For
instance, based on the index numbers fo r 2005 and 2008, we cannot say that prices rose
by 64.24% (216.56% — 152.32%) from 2005 to 2008. The actual percentage change from
2005 and 2008 is im plying that prices rose by 42.17%
from 2005 to 2008.
Alternatively, we can use index numbers directly to compare prices between 2005 and
2008 by making 2005 the base year. It may be more meaningful to compare 2008 val
ues w ith those in 2005 rather than the values in 2000. In fact, federal agencies routinely
update the base year used in their calculations o f statistical indices. For example, the
reported imports and exports price indices in 2008 have been updated from the base year
o f 1995 to a revised base year o f 2000.
It is fa irly simple to revise the base period o f an index. We basically transform the
index o f the newly chosen base period as 100 and values in other periods are adjusted by
the same proportion.
R E V I S I N G T H E BASE P E R I O D
A simple index can easily be updated with a revised base period as
EXAMPLE 19.8
Update the index numbers in Table 19.3 with a base year revised from 2000 to 2005.
SOLUTION: With a revised base of 2005, the index number for 2005 is updated
from 152.32 to 100. Other indices are adjusted according to the revision rule. For
instance, the index number for 2006 is updated as (171.52/152.32) X 100 = 112.61.
Table 19.4 contains index numbers that have been similarly updated.
TABLE 19.4 Price Index for Gasoline Using Base Year o f 2000 and 2005
Year 2000 2001 2002 2003 2004 2005 2006 2007 2008
Price 1.51 1.46 1.36 1.59 1.88 2.30 2.59 2.80 3.27
Price Index
(Base = 2000) 100 96.69 90.07 105.30 124.50 152.32 171.52 185.43 216.56
Price Index
(Base = 2005) 65.65 63.48 59.13 69.13 81.74 100.00 112.61 121.74 142.17
With the revised base of 2005, we can directly deduce that the gasoline price in
2008 was 142.17% of what it was in 2005, or 42.17% higher.
C A L C U L A T I O N OF AN U N W E I G H T E D A G G R E G A T E PRICE I N D E X
Let p it represent the price of item i in period t and let p i0 be the corresponding price
in the base period (t = 0). The unweighted aggregate price index in period t is
EXAMPLE 19.9
A real estate firm based in Florida collects data on the average selling price of con
dominiums, single-family homes, and multifamily homes that it sold over the last
three years. Table 19.5 shows the results. Compute the unweighted price index for
the properties, using 2007 as the base year.
TABLE 19.5 Average Price (in $1,000s) o f Properties Sold in Florida (Example 19.9)
Year Condominiums Single-fam ily M ultifam ily
2007 225 375 440
2008 148 250 390
2009 130 235 400
C A L C U L A T I O N OF A W E I G H T E D A G G R E G A T E PRICE
I N D E X : T H E L A S P E Y R E S P R IC E I N D E X
Let p it and qit represent the price and quantity o f item i in period t and let p i0 and
qi0 be the corresponding values in the base period (t = 0). Using only the base pe
riod quantities qi0, the Laspeyres price index for period t is
EXAMPLE 19.10
Table 19.6 shows the number o f condominiums, single-fam ily homes, and m u lti
fam ily homes sold in Florida. Use these quantities, along w ith the price inform ation
from Table 19.5, to compute the Laspeyres price index fo r real estate, given a base
year o f 2007.
S O L U T IO N : Since the Laspeyres price index evaluates the quantities in the base
period, we w ill only use the number o f properties sold in 2007 in the calculation.
Table 19.7 aids in the calculation o f the Laspeyres index.
Based on the above Laspeyres index, the real estate prices in 2008 were 69.90% o f
what they were in 2007, or equivalently they were 30.10% lower. Sim ilarly, the real
estate prices in 2009 were 33.80% lower. Note that the computed drop in property
values based on the Laspeyres price index is sharper than the one inferred from the
unweighted aggregate price index.
As mentioned earlier, the choice o f weights fo r the weighted aggregate price index de
pends on the quantity evaluated in a given period. Whereas the Laspeyres index uses the
base period quantities as weights, the Paasche index uses the current period quantities in
deriving the weights. Since the choice o f weights fo r the two methods are different, the
Laspeyres and Paasche indices d iffe r fo r the period under evaluation.
C A L C U L A T I O N OF A W E I G H T E D A G G R E G A T E P RI CE
I N D E X : T H E P A A S C H E P RI CE I N D E X
Let p it and qit represent the price and quantity o f item i in period t and let p i0 and qi0
be the corresponding values in the base period (t = 0). Using only the current period
(t = n ) quantities qin, the Paasche price index for period t is
EXAMPLE 19.11
Consider Tables 19.5 and 19.6, representing the price and quantity data for proper
ties sold in Florida. Use this inform ation to compute the Paasche price index fo r real
estate, given a base year o f 2007.
SOLUTION: Since the Paasche price index uses the quantities evaluated in the current
period, we use only the numbers of properties sold in 2009 in the calculations.
Table 19.8 aids in the calculation of the Paasche index.
The Paasche index is calculated as 68.80 for 2008 and 64.77 for 2009. Therefore,
according to the Paasche index with a base year of 2007, property values dropped
by 31.20% in 2008 and 35.23% in 2009.
In general, the Laspeyres and Paasche indices provide similar results if the periods being
compared are not too far apart. The two indices tend to differ when the length of time
between the periods increases since the relative quantities of items (weights) adjust to
the changes in consumer demand over time. Consumers tend to adjust their consumption
patterns by decreasing (increasing) the quantity of items that undergo a larger relative
price increase (decrease). For instance, a sharp increase in the price of an item is typically
accompanied by a decrease in the quantity demanded, making its relative weight go down
in value. Similarly, a sharp decrease in the price of an item will make its relative weight
go up. Therefore, a Paasche index that uses the updated weights theoretically produces a
lower estimate than a Laspeyres index when prices are increasing and a higher estimate
when prices are decreasing. Our results regarding property values are consistent with this
reasoning. During the period of price decline, the Laspeyres index suggests that relative
to 2007, property values have dropped by 30.10% and 33.80% in 2008 and 2009, respec
tively. According to the Paasche index for the same period, property values had larger
drops of 31.20% and 35.23%, respectively.
The Paasche index is attractive because it incorporates current expenditure patterns.
However, its data requirements are more stringent than those of the Laspeyres index. The
Paasche index requires that the weights be updated each year and the index numbers be
recomputed for all of the previous years. The additional cost required to process current
expenditure data, needed to revise the weights, can be substantial. It may not always be
possible to produce a timely Paasche index. Therefore, the Laspeyres index is a more
widely used weighted aggregate price index. The base period is changed periodically to
ensure that the Laspeyres index does not become outdated. Here the base period revision
involves updated calculations using quantity weights of the new base period.
EXAMPLE 19.12
Let us revisit the introductory case with the data presented in Table 19.1. Using
2007 as the base year, compute and interpret the weighted aggregate price indices
for liquor using
a. The Laspeyres method
b. The Paasche method
SOLUTION: Since 2007 is used as the base year, its value for both indices is set
equal to 100.
a. For the Laspeyres price index, the prices are weighted by the quantities evalu
ated in the base period of 2007. Therefore, the weighted price for 2007 is
computed as
Σ pit qi0 = 12.30 X 1,560 + 11.90 X 1,410 + 8.10 X 2,240 = $54,111.
Similarly, the weighted price equals
12.10 X 1,560 + 11.05 X 1,410 + 8.25 X 2,240 = $52,936.5 for 2008 and
9.95 X 1,560 + 10.60 X 1,410 + 7.95 X 2,240 = $48,276 for 2009.
The corresponding price index is (52,936.5/54,111) X 100 = 97.83 for
2008 and (48,276/54,111) X 100 = 89.22 for 2009. Therefore, based on the
Laspeyres index, liquor prices are 97.83% in 2008 and 89.22% in 2009 of
what they were in 2007. In other words, relative to 2007, overall liquor prices
dropped by 2.17% in 2008 and by 10.78% in 2009.
b. For the Paasche price index, the prices are weighted by the quantities evaluated
in the current period, which in our example is 2009. Therefore, the weighted
price for 2007 is computed as
Σ pit qin = 12.30 X 1,280 + 11.90 X 1,010 + 8.10 X 2,190 = $45,502.
SYNOPSIS OF I N T R O D U C T O R Y CASE
12. Consider the follow ing price and quantity data o f three a. Use 2004 as the base year to form a simple price index
products from 2008 to 2010. for tuition.
Year Product 1 Product 2 Product 3 b. Use 2007 as the base year to form a simple price index
Price $14.30 $13.90 $18.10 for tuition.
2008
Quantity 992 1,110 800 c. Compare the percentage tuitio n increase from 2004
Price $14.90 $13.70 $18.50 through 2007 and 2007 through 2009.
2009
Quantity 980 1220 790 17. F I L E According to dollar cost averaging, a fixed am ount o f
Price $15.50 $13.80 $17.90 money is invested periodically in a portfolio. Consequently,
2010
Quantity 140 1290 810 more units of a financial asset are purchased when prices
are low and fewer units are purchased when prices are high.
a. Compute the simple price index for each product, using Robert Dudek follows dollar cost averaging by making a
2008 as the base year.
m onthly investment o f $500 toward retirement. His m onthly
b. Compare the relative price movements o f the three investment is divided equally among tw o T. Rowe Price
products. mutual funds: Equity Income and Short-term Bond funds.
13. Use the price and quantity inform ation in the previous The follow ing table represents the m onthly adjusted close
exercise to com pute the follow ing aggregate price indices, price o f the funds in 2009. The data set can also be found on
given a base year o f 2008: the text website, labeled Returns 2 0 0 9 .
a. The unweighted aggregate price index M onth Eqlnc Bond M onth Eqlnc Bond
b. The Laspeyres price index January 14.77 4.47 July 18.4 4.72
c. The Paasche price index February 12.93 4.49 August 19.45 4.75
March 14.14 4.53 September 19.92 4.78
Applications April 16.04 4.58 October 19.53 4.80
14. Consider the follow ing average m onthly prices for regular May 16.87 4.64 November 20.60 4.85
gasoline in California in 2008. June 16.89 4.66 December 20.99 4.82
M onth Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec Source: http://finance.yahoo.com.
Price 3.25 3.18 3.56 3.82 3.97 4.48 4.46 4.16 3.79 3.39 2.46 1.82
a. Compute and interpret the Laspeyres price index.
Source: http://www.energyalmanac.ca.gov. b. Compute and interpret the Paasche price index.
a. Construct a simple price index w ith January 2008 as the c. Why are the results o f the tw o indices different?
base. 18. JJ Diner is a small mom and pop restaurant in Lincoln,
b. Determine the percentage change in the average Nebraska. They offer three choices for breakfast: omelets,
gasoline price in California from January to June. pancakes, or cereal. The average prices (in dollars) for
these options for 2007,2008, and 2009 are shown in the Region 2007 2008 2009
accompanying table. 271.5 240.7
Northeast 288.1
Year Omelet Pancakes Cereal
Midwest 161.4 150.5 142.5
2007 4.75 3.50 3.50
South 178.8 169.4 154.6
2008 5.25 4.25 4.00
West 342.5 276.1 224.2
2009 5.00 4.50 4.25
SOURCE: h ttp ://w w w .r e a lto r .o r g .
Use this inform ation, along w ith the price data provided in South 2,235 1,865 1,913
the previous exercise, to West 1,084 1,070 1,210
a. Compute and interpret the Laspeyres price index. SOURCE: h ttp ://w w w .r e a lto r .o r g .
N O M I N A L V E R S U S R EA L V A L U E S
A time series that has been deflated is said to be represented in real terms. The
unadjusted time series is said to be represented in nominal terms. We use a price
index to convert the nominal value of a time series into its real value as
Consider the following example. Lisa Redford has worked in a small marketing firm
in Florida for the last three years. Due to the recent economic crisis, her salary has
dropped from $80,000 in 2007 to $64,000 in 2009. While her salary has dropped by
20%, a larger drop in property values for the same time period may have made it easier
for Lisa to own a home in Florida. In Example 19.10, we used the base year of 2007 to
derive the Laspeyres price index of property values for 2009 as 66.20, implying that real
estate prices were 33.80% lower in 2009 than in 2007. It is more meaningful to compare
Lisa’s salary of $80,000 in 2007 (the base year) with the price-adjusted (real) salary of
($64,000/66.20) × 100 = $96,677 in 2009. Using the Laspeyres price index of prop
erty values for adjustment, Lisa is actually slightly better off in 2009 than she was in
2007, despite the salary cut. However, it is not reasonable to adjust Lisa’s salary solely
on the basis of the price index of property values in Florida. Since her expenditure is
not limited to mortgage payments, a more comprehensive price index is needed to make
the price adjustment to the salary. In fact, when we say that a series has been deflated,
we imply that the series has been adjusted on the basis of the price of a comprehensive
basket of goods and services.
The two most commonly used price indices used to deflate economic time series are
the Consumer Price Index, CPI, and the Producer Price Index, PPI. While both the
CPI and PPI measure the percentage price change over time for a fixed basket of goods
and services, they differ in the composition of the basket and in the types of prices used in
the analysis. The general process of computing the CPI and PPI is similar to the method
outlined in the preceding section. However, we will not elaborate on their composition in
this chapter.
The CPI is perhaps the best-know n weighted aggregate price index. The U.S.
Bureau of Labor Statistics com putes a monthly CPI based on the prices paid by
urban consum ers for a representative basket o f goods and services. As o f 2010, the
CPI uses 1982 as the base year. The prices o f several hundred consum ption items
are included in the index. In addition, random ly selected consum ers help determ ine
the expenditure for the representative basket of goods and services. The correspond
ing quantities of item s in the base year are used for com puting the weights for the
index.
The PPI is a weighted aggregate price index of prices measured at the wholesale, or
producer level. Prior to 1978, the PPI was called the Wholesale Price Index, WPI. The
U.S. Bureau of Labor Statistics computes a monthly PPI based on the selling prices
received by domestic producers for their entire marketed output. The target set includes
purchases of goods and services by consumers— directly from the producer or indi
rectly from a retailer— and by other producers as inputs to their production or as capital
investment.
Note that the CPI is based on out-of-pocket expenditures of an urban consumer
and the PPI is based on the portion that is actually received by the producer. There
fore, although sales and excise taxes are included in the CPI, they are not included
in the PPI because they do not represent revenue to the producer. The differences
between the PPI and CPI are consistent with the way these indices are used for defla
tion. It is common to use the CPI to adjust wages for changes in the cost of living.
The PPI, on the other hand, is useful to deflate revenue in order to obtain real growth
in output.
It is often assumed that the direction and magnitude of a price change in the PPI for
finished goods anticipates a similar change in the CPI for all items. This is not always
the case. In Figure 19.2, we plot the annual CPI and PPI from 1960-2010, with a base of
1982-1984 (Source: Bureau of Labor Statistics); the relevant data, labeled CPI & PPI,
are available on the text website. Interestingly, the two indices moved in sync until the
early 1980s. Beyond that, changes in prices that consumers paid far exceeded those re
ceived by producers, with the difference peaking in 2002. Also, noteworthy is the fact that
while there was a significant dip in the PPI, the CPI showed a very slight decline during
the peak of the financial crisis in 2009.
FIGURE 19.2 CPI and PPI for 1960-2010; base period 1982-1984
EXAMPLE 19.13
Tom Denio has been a project manager in a small construction firm in Atlanta since
2000. He started with a salary of $52,000, which grew to $84,000 in 2008. The
revenue of the construction firm also grew over the years, increasing from $13 mil
lion in 2000 to $18 million in 2008. According to the Bureau of Labor Statistics,
the values of the consumer price index with a base of 1982– 1984 for 2000 and 2008
are 172.20 and 215.30, respectively. The corresponding values of the producer price
index are 132.70 and 189.60, respectively.
a. Compute and analyze the nominal and real increase in Tom’s salary.
b. Compute and analyze the nominal and real revenue growth of the construction
firm.
SOLUTION:
a. Tom’s nominal salary grew by or by 62% from 2000 to
2008. This nominal salary makes no cost of living adjustment. We use the CPI
to compute his real salary as ($52,000/172.20) X 100 = $30,197 in 2000 and
($84,000/215.30) X 100 = $39,015 in 2008. These are Tom’s real salaries
based on 1982– 1984 prices. Thus, while Tom’s salary increased by 62% in dollar
amounts, his purchasing power increased by only or 29%.
b. The nominal revenue of the construction firm grew by or by
38% from 2000 to 2008. We use the producer price index to compute the
revenue growth in real terms. The real revenue is ($13/132.70) X 100 =
$9.80 million in 2000 and ($18/189.60) X 100 = $9.49 million in 2008.
Therefore, the real growth in revenue for the construction firm has been
Inflation Rate
The inflation rate is the percentage rate of change of a price index over time. We gener
ally use the CPI to compute the inflation rate in the U.S. Also, although it is common to
quote the inflation rate in annual terms, the CPI can be used to calculate the inflation rate
for any time period.
C A L C U L A T I N G T H E I N F L A T I O N RATE
The reported inflation rate it for a given period is generally based on the consumer
price index, CPI. It is computed as
EXAMPLE 1 9 .1 4
The consumer price indices for the years 2006, 2007, and 2008 are reported as
201.59, 207.34, and 215.30, respectively (Source: Bureau of Labor Statistics). Use
this information to compute the annual inflation rate for 2007 and 2008.
S O LU TIO N : The inflation rates for 2007 and 2008 are computed as:
Therefore, the inflation rate increased from 2.85% in 2007 to 3.84% in 2008.
EXAMPLE 19.15
At the beginning of 2007, Joe Gonzales invested $1,000 in a mutual fund, which
grew to $1,050 in a year. The consumer price index, with a base of 1982– 1984, is
203.37 for January 2007 and 212.23 for January 2008. Compute the real annual rate
of return for Joe.
SO LU TIO N : The real rate of return is based on the deflated investment values, which
for the two years are computed as
respectively. The resulting real return of investment is derived as
In Section 19.1, we used the Fisher equation to convert the nominal return into
the real return. The Fisher equation will give us this same value for the real return
on the investment. The nominal return for Joe is and the corre-
sponding inflation rate is Therefore, using the Fisher equa-
tion, we can compute to get the real rate of return
r = 0.0061, which varies slightly from the previous calculation for the real rate of
return due to rounding.
EXERCISES 19.3
Concepts 25. The follow ing table represents the nominal values o f an item
and the corresponding price index for 2007 and 2008.
22. The nominal values for four years are given by 32,37,39, and
42. Convert these values to real terms if the price index values Year Nominal Value Price Index
for the corresponding years are given by 100, 1 0 2, 103, 108. 2007 38 112
23. An item increases in value from 240 to 280 in one year. What 2008 40 120
is the percentage change in the value of this item? Compute
a. Compute the inflation rate for 2008.
the percentage change in real terms if overall prices have
increased by 5% for the same period. b. Compute the annual percentage change o f the item in
real terms.
24. Let revenues increase by 10% from $100,000 to $110,000.
Calculate the percentage change in real terms if the relevant 26. The follow ing table represents the nominal values o f an item
price index increases by 4% from 100 to 104. and the corresponding price index from 2009 to 2011.
Year Nominal Value Price Index Year 2001 2002 2003 2004 2005 2006 2007 2008 2009
2009 38 100 CPI 120.1 119.0 118.7 118.7 118.3 118.7 118.7 120.3 118.7
2010 40 103 Source: Bureau o f Labor Statistics.
2011 42 112 29. FILE Each m onth the Current Em ploym ent Statistics (CES)
program surveys numerous businesses and governm ent
a. Compare the percentage change in the nom inal
values w ith the corresponding real values from 2009 agencies in order to obtain detailed data on earnings
Month Sales PPI Month Sales PPI Use the following information on CPI and PPI for the
January 340,439 171.2 July 342,489 171.6 next three exercises.
February 342,356 170.9 August 350,800 174.1 Year CPI (1982-84 =100) PPI (1982 = 100)
March 339,228 169.6 September 343,687 173.3 2006 201.59 164.80
April 338,344 170.6 October 347,641 174.0 2007 207.34 172.70
May 339,873 170.6 November 354,467 176.6 2008 215.30 189.60
June 342,912 173.7 December 353,817 177.3 2009 214.54 172.90
a. How many tim es were nom inal sales below th a t o f the 30. The to ta l revenue fo r The W alt Disney Company was
previous month? $35,510,000 fo r 2007, $37,843,000 fo r 2008, and $36,149,000
b. Use the PPI to com pute the sales in real terms. How fo r 2009 (Source: http://finance.yahoo.com ).
many tim es were real sales below th a t o f the previous a. Deflate th e to ta l revenue w ith the relevant price index.
m onth? b. Discuss the revenue trend during the 2007–2009 period
c. Com pute the to ta l percentage increase in nom inal as using nom inal as w ell as real values.
w ell as real retail sales in 2009. 31. According to the New Hampshire D epartm ent o f Education,
d. Can econom ists feel o p tim istic about the economy the average teacher salary in public school districts in New
based on the retail sales data? Hampshire was $46,797 in 2006, $48,310 in 2007, and
28. Japan was the first Asian country to challenge the $46,797 in 2008. Com m ent on the percentage change in the
dom inance o f the U.S. in 1980s. However, since then, its do lla r value (nom inal) as w ell as the purchasing power (real)
York Times, O ctober 1 6 ,2010).This country has been trapped 32. According to Fisher College o f Business at the Ohio State
in low grow th and a dow nw ard spiral o f prices, known as University, the starting salary o f th e ir graduates in the MBA
deflation. Consider the fo llo w in g CPI o f Japan fo r the years program in 2008 was $89,156. W hat m ust be the starting
2001 through 2009. Com pute and in te rpre t the annual salary o f the MBAs in 2009 if th e salary increase makes the
in flatio n rates in Japan in the 2000s. exact cost o f living adjustm ent?
WRITING WITH STATISTICS
Ronald Wilson Reagan became the 40th President of the United States in 1981, after serv Sample
ing eight years as governor of California. He took office at a time when the U.S. was
experiencing economic stagnation and inflation. As president, Reagan advocated reduced Report—
business regulation and extensive tax cuts to boost economic growth. Arguably, the Reagan
era signifies a period of significant growth as the economy recovered from the recession. Economic
Crucial economic indicators were analyzed during Reagan’s presidency. The con
sumer price index (CPI) values imply that prices were 9.1% lower in 1981 and 24% Indicators
higher in 1989 than during the base years of 1982– 1984. The percentage price increase
during Reagan's term is calculated as 36.41%, resulting in an annualized inflation rate of during
(1 + 0.3641)1/8 — 1 = 3.96%. The CPI is also used to deflate crucial economic indicators.
For instance, while the median household income increased from $19,074 to $28,906, or Reagan's
by 51.55%, the corresponding deflated incomes increased from $20,984 to $23,311, or by
11.09%. Other similarly deflated economic indicators are presented in Table 19.A. Presidency
TABLE 1 9 .A Deflated Economic Indicators
Economic Indicators 1981 1989
Federal Debt ($ billions) $1,094.4 $2,312.9
Median Household Income $20,984 $23,311
Cost of a New Home $91,309 $ 120,000
The significant increase in the federal debt during the Reagan era is noteworthy. When
Reagan took office, he used deficit spending through tax cuts to stimulate the economy.
However, the debt continued to grow throughout the boom years. The resulting deflated
federal debt rose sharply from $1,094.4 billion in 1981 to $2,312.9 billion in 1989, or by
111%. The deflated cost of a new home grew from $91,309 to $120,000, or by 31.42%.
Therefore, despite the 11.09% growth in real income, a higher percentage increase in
home values made owning a new home more difficult. Interestingly, the deflated Dow
Jones Industrial Average High grew by a whopping 99.73% from 1,127 in 1981 to 2,251
in 1989. Finally, there was a steep decline of 40.79% in the deflated price of gasoline from
$ 1.52 per gallon to $0.90 per gallon. Perhaps the price decline was the consequence of the
falling demand as consumers reacted to the energy crisis of the 70s.
President Reagan’s policies reflected his personal belief in individual freedom. Ac
cording to Reagan supporters, his policies resulted in the largest peacetime economic
boom in American history. His critics, on the other hand, argue that the Reagan era is
associated with a widening of inequality, where the rich got richer with little economic
gains for most Americans. This argument is partly reflected by a meager 11.09% real
increase in the median household income during the supposedly good years.
Conceptual Review
LO 19.1 Define and com pute investment returns.
The investment return Rt is calculated as where and
are the capital gains yield and the income yield components, respectively.
The adjusted close prices make appropriate adjustments for dividend distributions,
stock splits and reverse stock splits. Let P*t and P*t-1 represent the adjusted close price
of a stock at times t (current) and t — 1 (prior), respectively. Using adjusted close prices,
the investment return Rt at the end of time t is calculated as
LO 1 9.2 Use the Fisher equation to convert nominal returns into real returns and vice versa.
LO 19.5 Compare the Laspeyres and the Paasche methods for com puting the w eighted
aggregate price index.
Let p it and qit represent the price and quantity of item i in period t and let p i0 and qi0
be the corresponding values in the base period (t = 0). Using only the base period
quantities qi0, the Laspeyres price index for period t is
Using only the current period (t = n) quantities qin, the Paasche price index for period
LO 1 9.6 Use price indices to deflate economic tim e series and derive the inflation rate.
A deflated time series is obtained by adjusting it for changes in prices, or infla
tion. A time series that has been deflated is said to be represented in real terms.
The unadjusted time series is said to be represented in nominal terms. We use a
price index to convert the nominal value of a time series into its real value as Real
Value
Two commonly used price indices used to deflate economic time series are the Con
sumer Price Index (CPI) and the Producer Price Index (PPI). It is common to use the
CPI to adjust wages for changes in the cost of living. On the other hand, the PPI is useful
to deflate revenue in order to obtain real growth in output. The reported inflation rate i,
for a given period is generally based on the CPI and is computed as
34. Toyota Motor Corp., once considered a company synonymous a. Compute and interpret the simple price index for each
with reliability and customer satisfaction, has been engulfed in a product, using 2009 as the base year.
perfect storm with millions o f cars recalled (BBCNews, March 19, b. Compute and interpret the unweighted aggregate price
2010).The following table shows the monthly adjusted close index, using 2009 as the base year.
price per share o f Toyota from October 2009 to March 2010. 37. Let the quantities corresponding to the prices in the previous
Adjusted Adjusted exercise be given by the follow ing table.
Date Close Price Date Close Price
Year Product 1 Product 2 Product 3
October 2009 78.89 January 2010 77.00
2009 90 32 48
November 2009 78.54 February 2010 74.83
2010 82 34 46
December 2009 84.16 March 2010 79.56
2011 76 30 36
Source h ttp ://fin a n ce .ya h o o .co m .
a. Compute the Laspeyres price index, using 2009 as the
a. Form a simple price index w ith October 2009 as the base. base year.
b. Update the simple price index, using January 2010 as the b. Compute the Paasche price index, using 2009 as the
base. base year.
c. What is the percentage price change from October 2009 c. Comment on the differences between the tw o indices.
to December 2009?
38. Lindsay Kelly bought 100 shares o f Google, 300 shares of
d. What is the percentage price change from January 2010
Microsoft, and 500 shares o f Nokia in January 2005. The
to March 2010?
adjusted close prices o f these stocks over the next three
35. Consider the follow ing price data from 2002 to 2010. years are shown in the accompanying table.
Year 2002 2003 2004 2005 2006 2007 2008 2009 2010 Year Google Microsoft Nokia
Price3.20 3.46 3.51 4.02 4.18 4.30 4.59 4.50 4.70 2005 195.62 24.11 13.36
2006 432.66 26.14 16.54
a. Compute the simple price index using 2002 as the base year.
2007 505.00 28.83 19.83
b. Update the index numbers w ith a base year revised from
Source: h ttp ://fin a n ce .ya h o o .co m .
2002 to 2005.
c. Plot the index numbers w ith a base year o f 2002 and a a. Compute and interpret the unweighted aggregate price
base year o f 2005. Compare the tw o plots. index for Lindsay's portfolio, using 2005 as the base year.
b. Com pute and in te rpre t the corresponding weighted a. Use the relevant price index to deflate the data on net
price index using the Laspeyres approach. revenue o f C itigroup, given in the previous exercise.
c. Why are the results from parts a and b so different? b. Use the relevant price index to deflate the data on net
39. Citigroup, Inc., is a m ajor financial services company based in incom e o f C itigroup, given in the previous exercise.
New York. It suffered huge losses during the global financial 41. An investor bought 1,000 shares o f C itigroup in January 2009
crisis and was rescued in November 2008 in a massive fo r $3.55 a share. She sold all o f her shares in December 2009
bailout by the U.S. governm ent. Consider the follow ing table, fo r $3.31 a share.
representing the net revenue and net income o f Citigroup for a. W hat annual rate o f return did the investor earn?
2006 to 2009. Both variables are measured in billions o f dollars. b. Use the CPI inform ation from the previous exercise to
Year Net Revenue Net Income com pute the in flatio n rate fo r 2009.
2006 146.6 21.2 c. W hat is the investor's real rate o f return?
2007 159.2 3.6
42. The adjusted close prices o f W endy's/Arby's Group, Inc., fo r
2008 105.8 -2 7 .7 the first three m onths o f 2008 are presented in the fo llo w in g
2009 111.0 - 1.6 table. Also included in the table is the corresponding
Source: h ttp ://m o n e y.cn n .co m . consumer price index (CPI).
a. Com pute and in te rpre t the sim ple price index fo r net Adjusted CPI (Base
Date Close Price 1982-1984)
revenue, using 2006 as the base year.
January, 2008 8.94 212.225
b. Compute and in te rpre t the sim ple price index fo r net
income, using 2006 as the base year. February, 2008 8.29 212.703
March, 2008 6.01 213.543
40. Consider the follo w in g consumer price index and producer
Source: http://finance.yahoo.com ; and Bureau o f Labor Statistics.
price index fo r 2006-2009.
Year CPI (1982-84 = 100) PPI (1982 = 100) a. Find the real rate o f return fo r the three m onths by first
2006 201.59 164.80 using the CPI to deflate the adjusted close price.
2007 207.34 172.70 b. Replicate the above result w ith Fisher's equation, based
2008 215.30 189.60 on th e nom inal rate o f return and the in flatio n rate.
2009 214.54 172.90
CASE STUDIES
Data fo r Case Study 19.1 M o n th ly Adjusted Close Prices fo r Four Firms, 1999–2000
FILE Month AMZN EBAY COKE JNJ
January, 1999 58.47 11.57 44.00 33.93
February, 1999 64.06 13.92 43.80 34.13
The hypothesis tests presented in earlier chapters make certain assumptions about the underlying
population. We refer to these tests as parametric tests. A t or an F test, for example, requires that
the observations come from a normal distribution. These tests are quite "robust," in the sense that
they are still useful when the assumptions are not exactly fulfilled, especially when the sample size
is large. However, in situations when the underlying population is markedly nonnormal, we apply
distribution-free alternative techniques called nonparametric methods. In addition to not needing
to fulfill a given distribution requirement, another benefit of nonparametric methods is that they do
not require a level of measurement as strong as that necessary for parametric tests. For instance, we
cannot calculate means and variances with ordinal data (required calculations for parametric tests)
because the numbers on an ordinal scale have no meaning except to indicate rank order. In this
chapter we explore a variety of nonparametric tests that make fewer assumptions about the distribu
tion of the underlying population and/or treat data of a weaker scale.
I N T R O D U C T O R Y C A S E
that is, the likelihood of extreme returns (area under the tail) is higher for a fatter-tailed
distribution than for a normal distribution. Figure 20.1 shows a normal distribution ver
sus a distribution with fatter tails. If Rebecca were to rely on tests that incorrectly assume
that the data are normally distributed, then there is a chance that she may make errone
ous conclusions. She chooses to use the Wilcoxon signed-rank test for the population
median.
Following the methodology outlined in earlier chapters, when conducting a hypoth
esis test for the population median m, we want to test whether m is not equal to, greater
than, or less than m0, the value of the population median postulated in the null hypothesis.
The null and alternative hypotheses will assume one of the following forms:
For the Metals fund return data in Table 20.1, we would like to determine whether the
median return for this fund is greater than 5%. We formulate the one-tailed test as
In order to arrive at the sample value for the Wilcoxon signed-rank test statistic T, several
calculations are necessary.
a. We first calculate the difference di between each observed value and the hypothe
sized median. In this case, di = x i — 5, as shown in the second column of Table 20.3.
TABLE 20.3 Calculations for the W ilcoxon Signed-Rank Test Statistic
Ranks of Negative Ranks of Positive
Return, x d = X − m0 |d | Rank Differences Differences
(1) (2) (3) (4) (5) (6)
− 7.34 − 7.34 − 5 = − 12.34 12.34 2 2
18.33 13.33 13.33 3 3
33.35 28.35 28.35 4 4
59.45 54.45 54.45 8 8
8.09 3.09 3.09 1 1
43.79 38.79 38.79 7 7
34.30 29.30 29.30 5 5
36.13 31.13 31.13 6 6
− 56.02 − 61.02 61.02 9 9
76.46 71.46 71.46 10 10
T+
44
T− = 11T
4
=
+
=
b. We then take the absolute value of each difference, | di|. See the third column
of Table 20.3. Any differences of zero are discarded from the sample; no zero-
differences occur in this example. We calculate | d i| because if the median is 5 (the
null hypothesis is true), then positive or negative differences of a given magnitude
are equally likely.
c. Next we rank the absolute value of each difference, assigning 1 to the smallest |di|
and n to the largest |di |. Note that n would be smaller than the original sample size if
there were some zero-difference observations, which we discarded; here, n equals the
original sample size of 10. Any ties in the ranks of differences are assigned the aver
age of the tied ranks. For instance, if two observations have the rank of 5 (occupying
the 5th and 6th positions), each is assigned the rank of (5 + 6 )/2 = 5.5. Or, if three
observations have a ranking of 1, each is assigned a rank of (1 + 2 + 3)/3 = 2. In
the Metals fund return example, there are no ties. The rankings for the differences are
shown in the fourth column of Table 20.3.
d. We then sum the ranks of the negative differences (denoted T )−and sum the ranks
of the positive differences (denoted T +). In this example we find two negative
differences, whose rank sum is T − = 11, and eight positive differences, whose rank
sum is T + = 44. These calculations are shown in the fifth and sixth columns of
Table 20.3.
The sum of T− and T + should equal n(n + l)/2 , which is the formula for the sum of
consecutive integers from 1 to n. In our example, T − + T + = 11 + 44 = 55. Also,
n(n + l) /2 = 10(10 + l ) / 2 = 55. As we will see shortly, the value of T is not used in
the actual analysis, but its calculation can help us avoid errors.
Values of T − and T + relatively close to one another indicate that the rank sums
more or less offset one another and provide evidence in support of the null hypoth
esis. However, a small value o f T − relative to T +, for example, implies larger posi
tive deviations from the hypothesized median value o f 5. This would suggest that the
median is greater than 5. We can view a small value o f T + relative to T− in a like
manner.
For ease o f exposition, we do not make a distinction between the random variable
and the particular outcom es o f the random variable in this chapter. For example,
we use the statistic T to represent a random variable as well as its sample value. We
adopt this same practice for the statistics W, H, rs, and R that we introduce in later
sections.
The critical value(s) for the Wilcoxon signed-rank test is found in Table 6 of
Appendix A; a portion of this table is shown in Table 20.4. A lower critical value TL is
used for a left-tailed test, an upper critical value TUis used for a right-tailed test, and both
lower and upper critical values are used for a two-tailed test.
TABLE 2 0 .4 Portion o f the Lower TL and Upper TU Critical Values for the Wilcoxon Signed-RankTest
Two-tailed Test: α = 0.10 α = 0.05 α = 0.02 a = 0.01
One-tailed Test: α = 0.05 α = 0.025 α = 0.01 a = 0.005
n= 8 5,31 3,33 0,36 —
9 8,37 5,40 1,44 0,45
10 10,45 8,47 3, 52 1,54
EXAMPLE 2 0.1
Given the return data in Table 20.1, determine whether the median return for the
Metals fund is significantly greater than 5% with α = 0.05.
SO LU TIO N : As discussed above, the competing hypotheses for the test are
H0: m ≤ 5 versus HA: m > 5. We already calculated the value of the test statistic as
T = T + = 44. For a right-tailed test with α = 0.05 and n = 10, the decision rule is to
reject the null hypothesis if the value of the test statistic T is greater than or equal to
TU = 45 (see the boldface value in Table 20.4). Since T = 44 < 45 = TU we do not
reject H0. At the 5% significance level, we cannot conclude that the median return is
greater than 5% for the Metals fund.
EXAMPLE 2 0 .2
Redo the test specified in Example 20.1 assuming that T is normally distributed.
SO LU TIO N :Again we use the competing hypotheses, H(): m ≤ 5 versus HA: m > 5,
and the value of the test statistic, T = T + = 44. Since there are 10 years of return data
(n = 10), we calculate the mean and standard deviation of the sampling distribution
of T as
We suggest that you take the time to show that a similar test employed on the Income fund
results in a test statistic value of T = T + = 40. Then, for α = 0.05, using the tests with
and without the normal distribution approximation, we cannot conclude that the median
return for the Income fund is greater than 5%.
1Since the normality assumption for param etric tests becomes less stringent in large samples, the main appeal of
rank-based tests tends to be with relatively small samples.
EXERCISES 20.1
Concepts Applications
1. Consider the follow ing com peting hypotheses and sample data. 6. During the fou rth quarter o f 2009, rents declined in alm ost
all m ajor cities in the United States. The largest fall was in
H0: m = 20 n = 7 T− = 2 T+ = 26
New York, where average rents fe ll nearly 20% to $44.69
Ha: m ≠ 20
per square fo o t annually (The Wall Street Journal, January 8,
a. Specify the decision rule at the 5% significance level.
2010). The fo llo w in g table lists the average rent per square
b. W hat is the conclusion? Explain.
fo o t for 10 cities during the fo u rth quarter o f 2009.
2. Consider the follow ing com peting hypotheses and sample data. Rent
City Rent City
H0: m ≥ 150 n = 9 T − = 42 T+ = 3 New York $45 Miami $24
HA: m < 150
Washington, D.C. 42 Seattle 24
a. Specify the decision rule at the 1% significance level.
San Francisco 30 Chicago 21
b. W hat is the conclusion? Explain.
Boston 30 Houston 20
3. Consider the follow ing com peting hypotheses and sample
Los Angeles 27 Philadelphia 20
data.
a. Specify the com peting hypotheses in order to determ ine
H0: m ≤ 1 50 n = 30 T− = 200 T + = 265
w hether the median rent is greater than $25 per square
HA: m > 150
foot.
a. Assuming th a t the sam pling d istributio n o f T is norm ally
b. Calculate the value o f the test statistic. Assume the
distributed, determ ine the value o f the test statistic.
norm al approxim ation for T.
b. Calculate the p -value.
c. Calculate the p-value.
c. At the 5% significance level, w hat is the conclusion?
d. At the 1% significance level, can you conclude th a t the
Explain.
median rent exceeds $25 per square foot?
4. Consider the fo llow ing sample data.
7. The follow ing table lists the annual returns (in percent) fo r a
8 5 11 7 6 5
top-perform ing m utual fund, ING Russia, over the last decade.
a. Specify the com peting hypotheses th a t determ ine Year Return Year Return
w hether the median is less than 10. 2000 − 17.80 2005 70.94
b. Calculate the value o f the test statistic T. 2001 80.32 2006 67.52
c. W ith α = 0.05, w hat is the decision rule? 2002 24.72 2007 30.69
d. Is the m edian less than 10? Explain. 2003 75.88 2008 − 71.51
5. Consider the follow ing sample data. 2004 5.91 2009 129.97
a. Specify the com peting hypotheses th a t determ ine a. Specify the com peting hypotheses in order to determ ine
w hether the m edian differs from 100. w hether the m edian return differs from 8%.
b. Calculate th e value o f the test statistic. Assume th a t the b. Calculate the value o f the test statistic. Do not assume
sam pling d istributio n o f T is norm ally distributed. the norm al approxim ation fo r T.
c. Determ ine the critical value(s) o f the test at α = 0.10. c. At the 10% significance level, w hat is the decision rule?
d. Is the m edian different from 100? Explain. d. Does the median return differ from 8 %?
Since applying the Wilcoxon signed-rank test to a matched-pairs sample is nearly iden
tical to its use for a single sample, we describe its use with an example and computer
output.
EXAMPLE 2 0 .3
We again use the Fund Returns data from the introductory case. At the 5% signifi
cance level, determine whether the median difference between the returns of the
Metals fund and the Income fund differs from zero.
SO LU TIO N : We must first recognize that these samples are dependent, in that each
return observation is blocked by year. We apply the Wilcoxon signed-rank test to
determine whether significant differences exist between the median difference of
the returns and formulate the two-tailed test as
Table 20.5 summarizes the method of calculating the value of the T statistic; that
is, we first calculate differences between the returns (column 4), find absolute
r + = 43
II
differences (column 5), and determine rankings (column 6). Then we compute the
sum of the ranks of negative differences (column 7, T − = 12), and the sum of the
ranks of positive differences (column 8, T + = 43). The value of the test statistic
T is T = T + = 43. Given the critical values in Table 20.4, the decision rule is to
reject H0 if T ≤ 8or T ≥ 47. Since 8 < T = 43 < 47, we do not reject H0; at the
5% significance level we cannot conclude that the median difference of the returns
differs from zero.
TABLE 20.7 Salary Inform a tion on C om puter Science and Finance Graduates (in $1,000s)
FILE Computer Science Finance
66 59 61 55
60 67 52 52
58 64 54 52
65 68 50 47
70 69 62 46
In order to determine whether salaries are significantly different depending on major, we
apply the Wilcoxon rank-sum test. Let m 1 and m2 denote the population median salary for
computer science and finance majors, respectively. We formulate the two-tailed test as
H0: m 1 — m2 = 0
HA: m 1 — m 2 # 0
In order to arrive at the value of the test statistic for the Wilcoxon rank-sum test W, several
steps are necessary.
a. We first pool the data from sample 1 (computer science) and sample 2 (finance),
with n 1 and n2 observations, and arrange all the data in ascending order of magni
tude. That is, we treat the independent samples as if they are one large sample of size
n1 + n2 = n. See column 1 of Table 20.8.
b. We then rank the observations from lowest to highest, assigning the numbers 1 to n.
Since we have a multiple tie at ranks 4, 5, and 6 , we assign to each of the tied obser
vations the mean of the ranks which they jointly occupy, or (4 + 5 + 6)/3 = 5. See
columns 2 and 3 of Table 20.8. We note that the finance salaries occupy the lower
ranks, whereas the computer science salaries occupy the higher ranks.
c. We then sum the ranks of the computer science salaries (denoted W1) and sum
the ranks of the finance salaries (denoted W2). Here we find that W1 = 149 and
W2 = 61; see columns 4 and 5 of Table 20.8. In order to check that we have
performed the calculations properly, we confirm that the sum of the rank sums,
which is equivalent to the sum of the integers
from 1 to n 1 + n2. We first find that W1 + W2 = 149 + 61 = 210. Since n1 = 10 and
n2 = 10, we then find that
61 Finance 12 12
62 Finance 13 13
64 Computer Science 14 14
65 Computer Science 15 15
66 Computer Science 16 16
67 Computer Science 17 17
68 Computer Science 18 18
69 Computer Science 19 19
70 Computer Science 20 20
W1= 149 W2= 61
If the median salary of computer science majors is equal to the median salary of finance
majors, then we would expect each major to produce about as many low ranks as high
ranks, so that W1is relatively close to W2. However, if the median salaries are significantly
different, then most of the higher ranks will be occupied by one major and most of the
lower ranks will be occupied by the other major, so that W 1will significantly differ from
W2. We determine whether W1is close to or far from W2 by comparing one of these values
to the appropriate critical value.
In the above example, since n 1 = n2 = 10, we select the value of the test statistic as the
rank sum of sample 1, so W = W1 = 149.
Table 20.9 shows a portion of lower WL and upper WUcritical values for the Wilcoxon
rank-sum test with n 1 and n2 as the number of observations in the respective samples. (A
more complete version is found in Table 7 of Appendix A.) Based on the specification of
the hypothesis test, the rejection region will be in either one or both sides.
TABLE 2 0.9 Lower WLand Upper WUC ritical Values fo r th e W ilcoxon Rank-Sum Test
a = 0.025 for one-tailed test and a = 0.05 for a two-tailed test
n,:3 4 5 6 7 8 9 10
n2: 8 8, 28 14, 38 21,49 29,61 39, 73 49,87 51,93 54,98
9 8,31 15,41 22,53 31,65 41,78 51,93 63,108 66,114
10 9,33 16,44 24,56 32, 70 43,83 54,98 66,114 79,131
EXAMPLE 2 0 .4
Use the salary data in Table 20.7 to determine whether the median computer science
salary differs from the median finance salary at the 5% significance level.
SOLUTION: As discussed earlier, the competing hypotheses for the test are
H0: m 1 - m2 = 0 versus HA: m 1 — m 2 ≠ 0 and the value of the test statistic is
W = W 1 = 149. For a two-tailed test with α = 0.05 and n 1 = n2 = 10, the decision
rule is to reject the null hypothesis if W ≤ 79 or W ≥ 131 (see the boldface values
in Table 20.9).
Since the value of the test statistic W = 149 is greater than 131, we can reject H0.
At the 5% significance level, we conclude that the median computer science salary
differs from the median finance salary.
Using a Normal D istribution Approxim ation for 1/1/
When «, and n2 both have at least 10 observations, we can use the normal distribution ap
proximation to implement a z test.
EXAMPLE 2 0 .5
Assuming that the distribution of W is approximately normal, let’s again determine
whether the median computer science salary differs from the median finance salary
at the 5% significance level.
Using the z table, we find the p -value as 2 X P ( Z ≥ 3.33) = 0.0009. Since the p -value
is less than the significance level of a = 0.05, we reject H 0 and conclude, as before,
that the median computer science salary differs from the median finance salary.
EXERCISES 20.2
b. Assuming th a t T is norm ally distributed, determ ine the 17. The fo llo w in g data are provided fo r tw o independent
value o f the test statistic. samples: W = 545, n1 = 25, and n2 = 25. Suppose the
c. Calculate the p-value. distrib u tio n o f W is approxim ately normal.
d. At the 1% significance level, w hat is the conclusion? a. Calculate the mean and standard deviation o f the
d istrib u tio n o f W.
12. Consider the fo llo w in g com peting hypotheses and
b. Specify the com peting hypotheses to determ ine
accom panying sample data drawn independently.
w hether the median o f Population 1 differs from the
H0: m 1 - m2 = 0 W1 = 80 W2 = 40
median o f Population 2.
HA: m 1 - m2 ≠ 0 n1 = 7 n2 = 8
c. Calculate the value o f th e test statistic Z.
a. Determ ine the value o f the test statistic.
d. Calculate the p-value.
b. Specify the decision rule at the 5% significance level.
e. At th e 10% significance level, w hat is the conclusion?
c. W hat is the conclusion?
19. A diet center claims that it has the most effective w eight loss c. W ith α = 0.05, w hat is the decision rule?
program in the region. Its advertisement says "Participants in d. Do the m edian test scores differ? Explain.
our program really lose weight." Five clients o f this program are 22 . F IL E According to the O rganization o f Economic
weighed on the first day o f the diet and then three months later. Cooperation and Developm ent, South Koreans spend
Weight on Weight Three m ore hours per year on the jo b than people in any other
Client First Day of Diet Months Later developed country (The Wall Street Journal, March 1,
1 155 151 2010). Suppose 10 workers in South Korea and 10 workers
2 205 203 in the United States are asked to report the num ber o f
hours w orked in the last year. The results are shown in the
3 167 168
accom panying table and also appear on the text website,
4 186 183
labeled South Koreans.
5 194 195
South Korea 2 6 2 4 1560 2698 2 730 2 8 7 9 3215 1753 2457 2669 2259
a. Specify the null and alternative hypotheses to test the United States 2132 1432 1718 1456 2323 1795 2861 1600 1104 1041
Kruskal-Wallis Test
An undergraduate admissions officer would like to examine whether SAT scores differ
by ethnic background. She collects a representative sample of SAT scores from Blacks,
Hispanics, Whites, and Asians. Her results are shown in Table 20.11. (The data are also
available on the text website, labeled KW-SAT.) She decides not to pursue the one-way
ANOVA F test because she does not believe that the population variances are equal. In
stead she chooses to apply the Kruskal-Wallis test.
Let m 1 m2, m3, and m4 denote the median SAT scores for Blacks, Hispanics, Whites, and
Asians, respectively. We formulate the competing hypotheses as
H0: m 1 = m2 = m3 = m4
HA: Not all population medians are equal
As in the Wilcoxon rank-sum test, we follow several steps to arrive at the Kruskal-Wallis
test statistic FI.
a. First, we pool the k independent samples (here, k = 4) and then rank the observa
tions from 1 to n. Since the total number of observations is 23, we rank the scores
from 1 to 23. As before, if there are any ties, then we assign to each of the tied
observations the mean of the ranks which they jointly occupy. In this sample, two
individuals score 1450 and each is assigned the rank of 12.5, since the values jointly
occupy the 12th and 13th ranks. Also, two individuals score 1800 and each is as
signed the rank of 20.5. Table 20.12 shows the rank for each SAT score.
b. We then calculate a ranked sum, denoted R h for each of the k samples. For instance,
the ranked sum for Blacks is calculated as 4 + 2 + 7 + 10 + 9 + 12.5 = 44.5.
These sums are shown in the second-to-last row of Table 20.12.
TABLE 20.12 C alculations fo r Kruskal-W allis Test
Blacks Rank Hispanics Rank Whites Rank Asians Rank
1246 4 1267 5 1581 15 1623 16
1148 2 1228 3 1649 18 1550 14
1300 7 1450 12.5 981 1 1936 23
1404 10 1351 8 1877 22 1800 20.5
1396 9 1280 6 1629 17 1750 19
1450 12.5 1800 20.5
1423 11
/?, = 44.5 R2 = 34.5 /?3 = 104.5 /?4 = 92.5
If median SAT scores across ethnic groups are the same, we expect the ranked sums R, to
be relatively close to one another. However, if these sums deviate substantially from one
another, then this is evidence that not all population medians are the same. We determine
whether the variability of the ranked sums significantly differs from one another by first
calculating the value of the test statistic H.
T H E T E S T S T A T I S T I C H F OR T H E K R U S K A L - W A L L I S T E S T
The test statistic H is defined as
where Ri and ni are the rank sum and the size of the ith sample, and k is the
number of populations (independent samples). If ni > 5 for i = 1 ,2 , . . . ,k , then H
can be assumed to follow the χ2 distribution with k - 1 degrees of freedom.
For small sample values (ni < 5), the test may be based on special tables; however, we
will not pursue that case. The Kruskal-Wallis test is always a right-tailed test.
EXAMPLE 2 0 .6
Use the data in Table 20.12 to conduct the Kruskal-Wallis test to determine if me
dian SAT scores differ by ethnic background at the 5% significance level.
S O LU TIO N: As discussed earlier, the appropriate hypotheses for the test are
H0: m 1 = m2 = m3 = m4
HA: Not all population medians are equal
We compute the value of the test statistic H as
EXAMPLE 2 0 .7
A sociologist suspects differences in median incomes in three major eastern cities. He
randomly samples 100 workers from each city. The top of Table 20.14 shows a portion
of the data; the full data set can be found on the text website, labeled City Income.
Unsure that the median income is the same in each city, he uses Minitab to conduct a
Kruskal-Wallis test and produces the results shown at the bottom of Table 20.14. Can
he conclude that median incomes differ at the 5% significance level?
85 63.3 80.1
Kruskal-WallisTeston Income
H = 85.50 DF = 2 P = 0.000 (adjusted for ties)
S O LU TIO N : We let m1, m2, and m3 denote the population median incomes for City 1,
City 2, and City 3, respectively, and formulate the competing hypotheses as
H0: m 1 = m2 = m3
HA: Not all population median incomes are equal
The value of the test statistic computed by Minitab is H = 85.50 with an associated
p-value ≈ 0. Since the p-value is less than the significance level of 0.05, we reject
H0 and conclude that not all median incomes across cities are the same.
EXERCISES 20.3
Concepts 26. FILE Four independent samples produced the results in the
23. Consider the fo llo w in g sample inform ation: k = 3 and accom panying table. The data are also available on the text
b. At the 10% significance level, w hat are the critical value Sample 3 4 6 8 10
c. Do the m edians differ? Explain. a. Specify the com peting hypotheses in order to test
24. Consider the follow ing sample inform ation: k = 5 and w hether differences exist between th e medians.
H = 12.4. b. W hat is the value o f the test statistic H?
a. Specify the com peting hypotheses in order to test c. Approxim ate the p-value.
w hether differences exist between the medians. d. A t the 1% significance level, do the medians differ? Explain.
b. Approxim ate the p-value.
Applications
c. At the 5% significance level, do the medians differ?
Explain. 27. F IL E A research analyst wants to test w hether the
unem ploym ent rate differs from one region o f the country
25. FILE Three independent samples produced the results in the
to another. She collects the unem ploym ent rate (in percent)
accom panying table. The data are also available on the te xt
o f similar-sized cities in three regions o f the United States.
website, labeled Exercise20.25.
The results are shown in the accompanying table, and
Sample 1 95 115 110 90
are also available on the te xt website, labeled Regional
Sample 2 | 85 105 80 75 90 Unemployment.
Sample 3 1 65 100 76 66 55 Region A 12.5 13.0 8.5 10.7 9.3
d. Does the unem ploym ent rate differ by region? Explain. three industries.
b. Calculate the value o f the test statistic H.
28. E
L
I
F A quality-control manager wants to test w hether there
is any difference in the length o f life o f three brands o f lig h t c. At the 10% significance level, w hat is the critical
375 280 350 compare the cleansing action o f the to p three detergents, 15
swatches o f w hite cloth were soiled w ith red w ine and grass
400 290 415
stains and then washed in front-loading machines w ith the
425 300 425
respective detergents. The follow ing whiteness readings are
410 325 380
shown in the accom panying table. The data are also on the
420 350 405
te xt website, labeled Detergents.
a. Specify the com peting hypotheses in order to test
Detergent 1 Detergent 2 Detergent 3
w hether differences exist in the median length o f life of
84 78 87
the three brands o f lig h t bulbs.
79 74 80
b. Calculate the value o f the test statistic H.
87 81 91
c. Approxim ate the p-value.
85 86 77
d. At the 10% significance level, do differences exist between
the median length o f life o f the three brands? Explain.
94 86 78
29. FILE A research analyst examines annual returns (in percent) a. Specify the com peting hypotheses in order to test
from Industry A, Industry B, and Industry C. The table below w hether differences exist in the cleansing action of
summarizes the relevant data w hich are also available on the the three detergents.
text website, labeled Industry Returns. b. Calculate the value o f the test statistic H.
Industry A Industry B Industry C c. Approxim ate the p-value.
16.86 15.41 13.53 d. At the 1% significance level, does cleansing action differ
5.11 10.87 9.58 by detergent? Explain.
Suppose we want to determine whether the observed relationship is real or due to chance.
As we noted before, return data often do not follow the normal distribution. Therefore,
using the t test to analyze the Pearson correlation coefficient is not appropriate. Instead,
we let ps denote the population Spearman rank correlation coefficient, and we formulate
a two-tailed test as
H0: ps = 0
Ha: ps ≠ 0
In order to conduct the test, we first calculate the sample Spearman rank correlation coef
ficient rs using the following steps.
a. We rank the observations from the Metals fund from smallest to largest. In the case
of ties, we assign to each tied observation the average of the ranks that they jointly
occupy. We perform the same procedure for the Income fund. Columns 2 and 3 of
Table 20.15 show the original return data, and columns 4 and 5 show the funds’
ranked values.
2d? = 76
II
o
b. We calculate the difference d, between the ranks of each pair of observations. See
column 6 of Table 20.15. As a check, when we sum the differences, 2 di, we should
obtain zero.
c. We then sum the squared differences. The resulting value is shown in the last cell of
column 7 in Table 20.15: E d2i = 76.
THE S P E A R M A N R AN K C O R R E L A T I O N TEST
The sample Spearman rank correlation coefficient rs between two variables x and
y is defined as
where di is the difference between the ranks of observations xi, and yi. If n ≤ 10, we
conduct the hypothesis test by comparing | rs| with positive critical values shown in
Table 8 of Appendix A.
If n > 10, rs can be assumed normally distributed with zero mean and standard
deviation of and hence the value of the resulting test statistic is computed as
We calculate rs as
A value of rs = 0.54 implies that the Metals and Income funds have a positive, rather
moderate, relationship.
For n ≤ 10, the rank correlation tests are based on special tables determined from the
exact distribution of rs. Table 20.16 shows a portion of upper critical values for one- and
two-tailed tests concerning ρ s at various a values. (A more complete version is found in
Table 8 of Appendix A.) For a two-tailed test with a = 0.05 and n = 10, the decision rule
is to reject the null hypothesis if | rs| > 0.648. See the relevant critical value in boldface
in Table 20.16.
TABLE 20.16 U pper C ritical Values fo r th e Spearman Rank C orrelation C oefficient
Two-tailed Test: α= 0.10 α = 0.05 α= 0.02 α = 0.01
One-tailed Test: α= 0.05 α = 0.025 α= 0.01 α = 0.005
n=8 0.643 0.738 0.833 0.881
9 0.600 0.683 0.783 0.833
10 0.564 0.648 0.745 0.794
Since the value of the test statistic, rs = 0.54, is not greater than 0.648, we cannot reject
H0. At the 5% significance level, we cannot conclude that the rank correlation between the
Metals fund and the Income fund differs from zero.
EXAMPLE 20.8
A sports statistician would like to analyze the relationship between a quarterback’s
salary (in $ millions) and his age. He collects data on 32 quarterbacks. The top of
Table 20.17 shows a portion of the data; the full data set can be found on the text web
site, labeled Quarterback Salary. He uses Minitab to calculate the Spearman rank
correlation coefficient and produces the result shown at the bottom of Table 20.17.
Can he conclude that salary and age are correlated at the 5% significance level?
F IL E Salary Age
25.5566 27
22.0441 26
0.6260 29
Spearman's rho = 0.375
SOLUTION: In order to determine whether salary and age are correlated, we
form ulate the competing hypotheses as
H0: ps = 0
HA: ps ≠ 0
Referencing Table 20.17, we find that rs = 0.375. Under normality, the value of the
corresponding test statistic is calculated as which yields
a p -value of 2 X P(Z ≥ 2.09) = 0.037. Since the p -value is less than the significance
level of 5%, we reject H0 and conclude that salary and age are correlated.
EXERCISES 20.4
Concepts 36. Consider the follow ing com peting hypotheses and
accom panying sample data.
31. Consider the follow ing com peting hypotheses and
accompanying sample data. H0: ρ s ≤ 0 rs = 0.64 and n = 50
Ha: ρ s > 0
H0: p s = 0 rs = 0.92 and n = 8
a. W hat is the value o f the test statistic and its associated
H a: P s ≠ 0
p-value? Assume the norm al approxim ation fo r rs.
a. W hat is the value o f the test statistic?
b. At the 1% significance level, w hat is the conclusion?
b. Specify the decision rule at the 1% significance level.
c. W hat is the conclusion?
Applications
32. Consider the fo llo w in g com peting hypotheses and
37. The fo llo w in g table shows the ranks given by tw o judges
accompanying sample data.
to the perform ance o f six finalists in a men's figure skating
H0: Ps ≥0 rs = - 0 .6 4 and n = 9 com petition:
H a: Ps < 0 Skater A B C D E F
a. W hat is the value o f the test statistic?
Judge 1 6 5 4 3 2 1
b. Specify the decision rule at the 5% significance level.
Judge 2 5 6 4 3 1 2
c. W hat is the conclusion?
a. Calculate and interpret the Spearman rank correlation
33. Consider the follow ing sample data:
coefficient rs.
x 12 18 20 22 25 15
b. Specify the com peting hypotheses in order to determ ine
y 15 20 25 22 27 19
w hether the correlation coefficient is significantly
a. Calculate and interpret rs. different from zero.
b. Specify the com peting hypotheses in order to determ ine c. At the 10% significance level, specify the critical value
w hether the Spearman population correlation coefficient and the decision rule.
differs from zero. d. W hat is the conclusion to the test? Are the ranks
c. At the 5% significance level, w hat is the decision rule? significantly related?
d. W hat is the conclusion? 38. FILEThe follow ing table shows the World Bank's 2008 ranking
34. Consider the follow ing sample data: o f the richest countries, as measured by per capita GNP. The
x 0 3 -1 4 7 data are also available on the text website, labeled WB ranking.
-2
y In addition, it gives each country's respective rank w ith respect
-4 -3 -8 -5 -9 -1 0
to infant m ortality according to the Central Intelligence Agency.
a. Calculate and interpret rs. A higher rank indicates a lower m ortality rate.
b. Specify the com peting hypotheses in order to determ ine Country Per Capita GNP Rank Infant Mortality Rank
w hether the Spearman population correlation coefficient Luxembourg 1 7
is less than zero.
Norway 2 4
c. At the 1% significance level, w hat is the decision rule?
Singapore 3 1
d. W hat is the conclusion?
United States 4 10
35. Consider the fo llow ing com peting hypotheses and Ireland 5 9
accom panying sample data.
Switzerland 6 5
H0: ρ s = 0 rs = -0 .8 5 and n = 65 Netherlands 7 8
Ha: ρ s ≠ 0
Austria 8 6
a. W hat is the value o f the test statistic and its associated
Sweden 9 2
p-value? Assume the normal approxim ation for rs.
Iceland 10 3
b. At the 10% significance level, w hat is the conclusion?
a. Calculate and interpret the Spearman rank correlation c. At the 5% significance level, specify the critical value
coefficient rs. and the decision rule.
b. Specify the com peting hypotheses in order to determ ine d. Are these variables related? Explain.
w hether the correlation coefficient is significantly 41. FILE The dire cto r o f graduate adm issions at a local
diffe re nt from zero. un ive rsity is analyzing th e relation ship betw een scores on
c. At the 5% significance level, specify the critical value and the G raduate Record Exam ination (GRE) and subsequent
the decision rule. perform ance in graduate school, as measured by a
d. Are GNP and the infant m o rta lity rate correlated? student's grade p o in t average (GPA). She uses a sample
Explain. o f 7 students w ho graduated w ith in th e past five years.
The data are in th e accom panying table and on th e te xt
39. You are interested in w hether the returns on Asset A
w ebsite, labeled GRE/GPA.
are negatively correlated w ith the returns on Asset B.
Consider the fo llow ing annual return data on the tw o Student 1 2 3 4 5 6 7
assets: GRE 1500 1400 1000 1050 1100 1250 800
Asset A Asset B GPA 3.5 3.0 2.9 3.0 3.3 2.7
4
.
3
Year 1 - 20 % 8%
a. Calculate and interpret the Spearman rank correlation
Year 2 -5 5 coefficient rs.
Year 3 18 -1 b. Specify the com peting hypotheses in order to
Year 4 15 -2 determ ine w hether GRE and GPA are positively
related.
Year 5 4 3
c. At the 5% significance level, specify the critical value and
Year 6 -12 2
the decision rule.
a. Calculate and interpret the Spearman rank correlation d. Are GRE and GPA positively related? Explain.
coefficient rs.
42. A social scientist analyzes th e relation ship betw een
b. Specify the com peting hypotheses in order to educational a tta in m e n t and salary. For 65 individuals
determ ine w hether the asset returns are negatively he collects data on each in dividu al's educational
correlated. a tta in m e n t (in years) and h is/h er salary (in $ 1,0 0 0 s).
c. At the 1% significance level, specify the critical value. He the n calculates a Spearman rank correlation co e fficien t
d. W hat is the conclusion to the test? Are the returns o f 0.85.
negatively correlated? a. Specify the com peting hypotheses in order to
40. FILE In an attem p t to determ ine w hether a relationship determ ine w hether educational attainm en t and
exists between the price o f a home and the num ber o f days salary are related.
it takes to sell the home, a real estate agent collected the b. Assume th a t the d istributio n o f rs is approxim ately
fo llow ing data from recent sales in his city (the data are also normal. Calculate the value o f the test statistic and the
available on the te xt website, labeled Price/Days): p-value o f the test.
Price (in $1,000s) Days to Sell Home c. At the 5% significance level, are educational attainm ent
and salary related?
265 136
43. An engineer examines the relationship between the w e igh t
225 125
o f a car and its average miles per gallon (MPG). For a sample
160 120
o f 100 cars, he calculates a Spearman rank correlation
325 140 coefficient o f —0.60.
430 145 a. Specify the com peting hypotheses in order to
Excellent 4 Fair 2
Good 3 Poor 1
Here the value attached to excellent (4) is higher than the value attached to good (3), indi
cating that the response of excellent is preferred to good. However, another representation
of the ratings might be
Excellent still receives a higher value than good, but now the difference between the two
categories is 20 (100-80), as compared to a difference of 1 (4-3) when we use the first
classification. In other words, differences between categories are meaningless with ordi
nal data.
We use the sign test on a matched-pairs sample of ordinal data to determine whether
there are significant ordinal differences in the population. When applying the sign test, we
are only interested in whether the difference between two values in a pair is greater than,
equal to, or less than zero. The difference between each pairing is replaced by a plus sign
(+ ) if the difference is positive (that is, the first value exceeds the second value) or by a
minus sign (-) if the difference between the pair is negative. If the difference between the
pair is zero, we discard that particular observation from the sample.
If significant differences do not exist between the two populations, then we expect just
as many plus signs as minus signs. Equivalently, we should observe plus signs 50% of the
time and minus signs 50% of the time. Suppose we let p denote the population proportion
of plus signs. (We could just as easily allow p to represent the population proportion of
minus signs without loss of generality.) The competing hypotheses for the sign test take
one of the following forms.
A two-tailed test allows us to determine whether the proportion of plus signs differs
from the proportion of minus signs. A right-tailed (left-tailed) test allows us to determine
whether the proportion of plus signs is greater than (less than) the proportion of minus
signs.
Let be the estimator of the population proportion of plus signs. As discussed
in Chapter 7, if np and n( 1 — p) are both 5 or more, then the distribution of is ap-
proximately normal, with mean E(P) = p and standard deviation
Assuming a probability of success p = 0.50, the approximation is satisfactory as long as
n ≥ 10. When n < 10, we rely on the binomial distribution to conduct the sign test; we
will not consider such cases.
T H E T E S T S T A T I S T I C F OR T H E S I G N T E S T
The test statistic is assumed to follow the z distribution, and its value is computed
as where p = x /n is the sample proportion of plus signs. The test is
valid when n ≥ 10.
EXAMPLE 2 0 .9
In December 2009, Domino’s Pizza, Inc. released untraditional ads citing that its
old recipe for pizza produced crust that tasted like cardboard and sauce that tasted
like ketchup. Domino’s Pizza claims that its reformulated pizza is a vast improve
ment over the old recipe; for instance, garlic and parsley are now baked into the
crust and a new sweeter, bolder tomato sauce is used. Suppose 20 customers are
asked to sample the old recipe and then sample the new recipe. Each person is
asked to rate the pizzas on a 5-point scale, where 1 = inedible and 5 = very tasty.
The ratings are shown in Table 20.19; the data labeled D o m in o ’s Pizza can also be
found on the text website. Do these data provide sufficient evidence to allow us to
conclude that the new recipe is preferred to the old recipe? Use α = 0.05.
2 3 2 + 12 4 5 -
3 2 5 - 13 1 2 -
4 4 4 0 14 3 3 0
5 2 5 - 15 5 3 +
6 1 3 - 16 3 4 -
7 3 2 + 17 1 5 -
8 1 2 - 18 4 2 +
9 2 4 - 19 3 4 -
10 4 5 - 20 2 5 -
SO LU TIO N : If customers feel that there is no difference between the old recipe and the
new recipe, then we expect 50% of the customers to prefer the old recipe and 50% to
prefer the new recipe. Let p denote the population proportion of consumers that prefer
the old recipe. We want to specify the competing hypotheses such that rejection of the
null hypothesis provides evidence that customers prefer the new recipe (implying that
p is significantly less than 0.50). We set up the competing hypotheses as
H0: p ≥ 0.50
HA: p < 0.50
Table 20.19 shows the signs for each customer. For example, customer 1 ranks the
old recipe with the value 3 and the new recipe with the value 4, which yields a minus
sign: 3 — 4 = —1. This difference implies that this customer prefers the new recipe.
We find 4 positive signs, 14 negative signs, and 2 ties (ranks of zero). We then let n
denote the number of matched-paired observations such that the sign between the
rankings is nonzero; thus, n equals 18. We denote as the sample proportion of plus
signs. Given that there are four plus signs, the sample proportion is calculated as
(Note that if we had calculated the sample proportion of minus
signs, then the resulting value of the test statistic only differs in its sign.
In this instance, we would conduct a right-tailed hypothesis test.) We calculate the
value of the test statistic as
Using the z table, we find the p-value for a left-tailed test as P(Z ≤ -2 .3 7 ) =
0.0089. Since the p-value is less than the significance level of α = 0.05, we reject
H0 and conclude that consumers prefer the reformulated version as compared to the
old recipe at the 5% significance level.
The sign test can be used with quantitative as well as ordinal data. However, since the sign
test ignores the magnitude in the difference between two observations, it is advisable to
use the Wilcoxon signed-rank test if quantitative data are available.
EXERCISES 20.5
Concepts Applications
44. Consider the follow ing com peting hypotheses and sample FILEConcerned w ith the increase o f plastic water bottles in
48. ____
data. landfills, a leading environm entalist wants to determine whether
H0:p = 0.50 n = 40 there is any difference in taste between the local tap water and
HA: p ≠ 0.50 the leading bottled water. She randomly selects 14 consumers
a. Determ ine the value o f the test statistic. and conducts a blind taste test. She asks the consumers to rank
b. Calculate the p-value. the taste on a scale of one to five (with "five" indicating excellent
taste). The sample results are shown in the accompanying table
c. At the 5% significance level, w hat is the conclusion?
Explain. and are also available on the text website, labeled Water.
Tap Bottled Tap Bottled
45. Consider the follow ing com peting hypotheses and sample data.
Consumer Water Water Consumer Water Water
H0: p ≤ 0.50 n = 25
HA: p > 0.50 1 4 5 8 5 2
a. Determ ine the value o f the test statistic. 2 3 2 9 3 4
a. Specify the com peting hypotheses to determ ine a. Specify the com peting hypotheses to determ ine w hether
w hether the proportion o f negative signs differs from the there are significant differences in preferences between
proportion o f positive signs. tap w ater and bottled water.
b. Calculate the value o f the test statistic. b. Calculate the value o f the test statistic.
d. At the 5% significance level, w hat is the conclusion? Explain. d. At the 5% significance level, w hat is the conclusion? Do
the results indicate th a t significant differences exist in
47. Consider the follow ing sign data, produced from a m atched-
preferences?
pairs sample o f ordinal data.
49. In March 2 0 0 9 , 100 registered voters were asked to rate the
"effectiveness" o f President Obama. In March 2010, these
a. Specify the com peting hypotheses to determ ine w hether same people were again asked to make the same assessment.
the proportion o f negative signs is significantly greater Seventy percent o f the second ratings were lower than the
than the proportion o f positive signs. first ratings and 30% were higher.
b. Calculate the value o f the test statistic. a. Specify the com peting hypotheses to determ ine w hether
c. W hat is the decision rule w ith a 1% level o f significance? the President's rating has significantly declined.
d. W hat is the conclusion? Explain. b. Calculate the value o f the test statistic.
c. Calculate the p-value. b. Calculate the value o f the test statistic.
d. At the 5% significance level, do the data suggest tha t the c. Determine the critical values at the 10% significance level.
President's rating has significantly declined? d. Do the data suggest significantly different faculty
50. FILE For scholarship purposes, tw o graduate faculty ratings?
members rate 12 applicants to the PhD program on a 51. A new die t and exercise program claims th a t it significantly
scale o f 1 to 10 (w ith 10 indicating an excellent candidate). lowers a participant's cholesterol level. In order to test this
These ratings are shown in the fo llow ing table and are also claim, a sample o f 60 participants is take n. Their cholesterol
available on the te xt website, labeled PhD Rating. levels are measured before and after the three-m onth
Faculty Faculty Faculty Faculty program. Forty o f the participants recorded lower cholesterol
Candidate A's Rating B's Rating Candidate A's Rating B's Rating levels at the end o f the program , 18 participants recorded
Sequence: OOOOUUUOOOOUOOOUUUUOOOOUUOOOOO
One possible way to test whether or not a machine is operating properly is to determine
if the elements of a particular sequence of Os and Us occur randomly. If we observe a
long series of consecutive Os (or Us), then the machine may be significantly overfilling
(or underfilling) the cereal boxes. Adjustment of the machine is likely necessary if this
is the case. Given the observed sequence, can we conclude that the machine is operating
properly in the sense that the series of Os and Us occur randomly?
In general, when applying the runs test, we specify the competing hypotheses as
H0: The elements occur randomly.
HA: The elements do not occur randomly.
In this particular application, the null hypothesis implies that the machine properly fills
the boxes and the alternative hypothesis implies that it does not. Before deriving the
test statistic, it is first necessary to introduce some terminology. We define a run as an
uninterrupted sequence of one letter, symbol, or attribute, such as O or U. We rewrite
the observed sequence, but now include single horizontal lines below the letter O. The
five single lines indicate that we observe five runs of O, or R0 = 5. Similarly, the double
horizontal lines below the letter U show that we have four runs of U, or R U = 4. Thus,
the total number of runs R is equal to nine: R = R0 + RU = 5 + 4 = 9. Also, note that
we have a total of 30 observations, of which 20 are Os and 10 are Us, or n = n0 + nU =
20 + 10 = 30.
We then ask: “Are nine runs consisting of 30 observations too few or too many compared
with the number of runs expected in a strictly random sequence of 30 observations?”
In general, the runs test is a two-tailed test; that is, too many runs are deemed just as
unlikely as too few runs. For example, consider the following two sequences:
If the null hypothesis of randomness is true, Sequence A seems unlikely in the sense
that there appear to be too few runs given a sample of 30 observations. Sequence B also
seems unlikely since O and U alternate systematically, or equivalently, there appear to
be too many runs. It is more readily apparent in the machine-filling application that a
sequence that produces too few runs indicates a machine that is not operating prop
erly; that is, the machine has a pattern of consistently overfilling and/or underfilling
the cereal boxes. However, a machine that exhibits a perfect regularity of overfilling,
underfilling, overfilling, underfilling, etc. (too many runs) may be just as problematic.
If there are too many runs, then this may indicate some sort of repeated alternating
pattern.
Let n1 and n2 denote the numbers of Os and Us in an n-element sequence. In general,
the sampling distribution of R (the distribution for the runs test) is quite complex and its
critical values are provided in specially constructed tables. However, if n 1 and n2 are at
least 10, then the distribution of R is approximately normal.
T H E T E S T S T A T I S T I C R F OR T H E W A L D - W O L F O W I T Z R U N S T E S T
The test statistic is assumed to follow the z distribution, and its value is computed
where R represents the number of runs with mean
and standard deviation n1 and n2 are the number of elements
in a sequence possessing and not possessing a certain attribute and n = n 1 + n2. The
test is valid when n 1 ≥ 10 and n2 ≥ 1 0 .
Thus, the expected number of runs in a sample with 30 observations is 14.33 and the
standard deviation is 2.38. We calculate the value of the test statistic as
We find the p -value for a two-tailed test as 2 X P(Z < -2 .2 4 ) =
0.0250. Since the p -value is less than α = 0.05, we reject H0 and conclude that the
machine does not properly fill the boxes. At the 5% significance level, adjustment of the
machine is necessary.
The M eth o d of Runs Above and Below the M edian
As mentioned earlier, the runs test can be applied to both qualitative and quantitative data.
Any sample with numerical values can be treated similarly by using letters, say A and B,
to denote values falling above and below the median of the sample, respectively. We omit
values that are equal to the median. The resulting As and Bs can be tested for randomness
by applying the method of runs above and below the median. This test is especially
useful in detecting trends and cyclical patterns in economic data. A finding of too few
runs is suggestive of a trend; that is, we first observe mostly As and later mostly Bs (or
vice versa). In computing the value of the test statistic, we omit values that are equal to
the median. A systematic alternation of As and Bs— that is, too many runs— implies a
cyclical pattern. Consider the following example.
EXAMPLE 2 0 .1 0
Table 20.20 shows the growth rate in the gross domestic product (GDP) for the
United States from 1980 through 2009; the data labeled US GDP can also be found
on the text website. Use the method of runs above and below the median with a
significance level of 10% to test the null hypothesis of randomness against the alter
native that a trend or cyclical pattern occurs.
SO LU TIO N : Since we are testing the null hypothesis of randomness against the
alternative that there is a trend or cyclical pattern, we formulate the competing
hypotheses as
H0: The GDP growth rate is random.
HA: The GDP growth rate is not random.
We first calculate the median GDP growth rate as 3.21%. Letting A and B denote
an observation that falls above the median and below the median, respectively, we
rewrite the data using the following sequence of As and Bs:
We see that the number of runs below the median RB is 6, while the number of runs
above the median RA is 5, so the total number of runs R is 11. Also, since no values
were discarded (no value was equal to the median), the total number of observations
are n = 30, where the number of observations below the median and the number of
observations above the median are nB = 15 and nA = 15, respectively. Using this in
formation, we compute the mean and the standard deviation of the distribution of R as
EXERCISES 20.6
Concepts 53. Consider the follow ing inform ation: n 1 = 10, n2 = 13 and
R = 8, where R is the num ber o f runs, n] and n2are the
52. Consider the follow ing inform ation: n1 = 24, n2 = 28 and
num ber o f elements in a sequence possessing and not
R = 18, where R is the num ber o f runs, n 1and n2are the
possessing a certain attribute, and n 1 + n2 = n.
number o f elements in a sequence possessing and not
possessing a certain attribute, and n 1 + n2 = n. a. Specify the com peting hypotheses in order to test for
nonrandomness.
a. Specify the com peting hypotheses in order to test for
b. What is the value o f the test statistic?
nonrandomness.
b. What is the value o f the test statistic? c. Calculate the p -value.
DDDUUDUUUUUDDDUUDUUUDDDUUUUDDD
59. ranThe s u p e rin te n d e n t o f a large suburban high school
m ust decide w h e th e r to close th e school fo r at least tw o
At the 1% significance level, conduct a hypothesis test in days due to th e spread o f flu . If she can con firm a tre n d
order to determ ine if the outcom es are nonrandom .
in absenteeism , the n she w ill close th e high school. The
fo llo w in g are th e num ber o f students absent from th e
Applications
high school on 25 consecutive school days (the data are
56. Given the digits zero throu gh nine, a com puter program is
also available on th e te x t w e bsite labeled Absenteeism):
supposed to generate even and odd num bers random ly.The
4 4 , 5 6 , 5 5 , 4 0 , 4 2 , 5 1 , 5 0 , 5 9 , 5 8 , 4 5 , 4 4 , 5 2 , 5 2 , 4 3 , 4 8 , 5 8 , 57,
com puter produced the fo llo w in g sequence o f numbers:
24 , 6 0 , 6 5 , 6 9 , 7 5 , 7 0 , 72, 72
5346802977168315243392
Use the m ethod o f runs above and below the median and
a. Specify the com peting hypotheses in order to test for
α= 0.05 to test the null hypothesis o f randomness against
nonrandomness.
the alternative th a t there is a trend.
b. W hat is the value o f the test statistic?
60. FILE A research analyst follow s the biotechnology industry
c. W hat is the decision rule w ith a 1% level o f significance?
and examines the daily stock price o f Amgen, Inc. over the
d. Is the program operating properly? Explain.
past year. The table below shows a portion o f the daily stock
57. A gam bler suspects th a t a coin may be w eighted more price o f Amgen for the 252 trading days in 2010; the fu ll data
heavily tow ard the outcom e o f tails (T) over heads (H). He set can be found on the te xt website, labeled Amgen. The
flips the coin 25 tim es and notes the follow ing sequence: research analyst wants to test the random -w alk hypothesis
TTHTTTHHTHTTHTTTHHTHTHTTH th a t suggests th a t stock prices move random ly ove rtim e
a. Specify the com peting hypotheses in order to test the w ith no discernible pattern.
Meg Suzuki manages a trendy sushi restaurant in Chicago, Illinois. She is planning an
aggressive advertising campaign to offset the loss of business due to competition from
other restaurants. She knows advertising costs increase overall costs, but she hopes this
effort will positively affect sales, as it has done in the past under her tenure. She collects
monthly data on sales (in $ 1,000s) and advertising costs (in $) over the past two years and
produces the following regression equation:
Estimated Sales = 17.77 + 0.03Advertising Costs
t-statistics = (17.77) (21.07)
At the 5% significance level, Meg initially concludes that adver
tising is significant in explaining sales. However, to estimate this
regression model, she had to make certain assumptions that might
not be valid. Specifically, with a time series analysis, the assump
tion maintaining the independence of the error terms often breaks
down. In other words, the regression model often suffers from
serial correlation. Table 20.22 shows a portion of the values of the
residuals; the full data set can be found on the text website, labeled
Residuals.
TA B LE 2 0 . 2 2 Values o f Residuals
Observation Residual
1 -0 .3 1
2 - 0 .8 0
⋮ ⋮
24 0.12
Meg would like to use the runs test to determine whether the positive and negative residu
als occur randomly at the 5% significance level.
One of the underlying assumptions of a linear regression model is that the error term is Sample
uncorrelated across observations; that is, there is no serial correlation. In a regression
model relating sales to advertising costs, there is reason to believe that serial correlation Report—
may be a problem because the data are time series. Figure 20.A is a scatterplot of the
residuals against time. If the residuals show no pattern around the horizontal axis, then Testing the
serial correlation is not likely a problem. Given the wavelike movement in the residuals
over time (clustering below the horizontal axis, then above the horizontal axis, etc.), posi Independence
tive serial correlation is likely a problem in this application.
of Residuals
The above graphical analysis is supplemented with a runs test to determine if the residuals
fail to follow a random pattern. A residual is given a + symbol if the residual is positive
and a – symbol if the residual is negative. There are 12 positive residuals and 12 negative
residuals, or n + = 12 and n _ = 12, respectively. A run is then defined as an uninterrupted
sequence of a + or a - sign. The sample data exhibits three positive runs, R + = 3, and
three negative runs, R- = 3, for a total number of runs equal to six, R = 6.
Are six runs consisting of 24 observations too few or too many compared with the
number of runs expected in a strictly random sequence of 24 observations? In order to
answer this question, the mean and the standard deviation for the distribution of R are
calculated. The mean number of runs in a sample of 24 observations is 13 with a standard
deviation of 2.4. Table 20.A provides summary data to conduct the runs test.
The sample value of the test statistic is z = − 2.92 with an associated p -value of 0.0036.
The null hypothesis of the randomness of the residuals is rejected at the 5% level; the
pattern of the residuals is nonrandom. Moreover, given that the actual number of runs is
significantly less than the expected number of runs, or too few runs, positive serial cor
relation is likely to be the problem. Corrective measures should be taken before statistical
inference is conducted on the estimated model.
Conceptual Review
LO 20.1 Distinguish between parametric and nonparametric tests.
Nonparametric tests, also referred to as distribution-free tests, do not require stringent
assumptions of parametric tests and are especially attractive when the underlying popu
lation is markedly nonnormal. Also, while parametric tests require data of interval or
ratio scale, nonparametric tests can be performed on data of nominal or ordinal scale.
However, a nonparametric test ignores useful information since it often focuses on
the rank rather than the magnitude of sample values. Therefore, in situations when the
parametric assumptions are valid, the nonparametric test is less powerful (more prone
to Type II error) than its parametric counterpart. In general, when the assumptions for
a parametric test are met, it is preferable to use a parametric test rather than a non
parametric test. Since the normality assumption for parametric tests is less stringent
in large samples, the main appeal of rank-based nonparametric tests tends to be with
relatively small samples.
LO 20.3 Determine whether the population median difference differs from zero under
matched-pairs sampling.
We can also apply the Wilcoxon signed-rank test as the nonparametric counterpart
to the t test that was used to determine whether two population means differ under
matched-pairs sampling. The measurement of interest is the difference between paired
observations, or d, = x i — yi . We conduct the test by following analogous steps to those
applied for a one-sample Wilcoxon signed-rank test.
LO 20.4 Determine whether two population medians differ under independent sampling.
We use the Wilcoxon rank-sum test to determine whether two populations have dif
ferent medians under independent sampling. We pool the data and calculate the rank
sum of sample 1, W1 and the rank sum of sample 2, W2. If the two sample sizes satisfy
n 1 ≤ n2, then the test statistic W is W = W1. Otherwise, we set W = W2.
If either sample is less than or equal to 10, we use a special table to obtain the
critical value(s) for the hypothesis test. The sampling distribution of W can be
approximated by the normal distribution if both sample sizes are greater than
or equal to 10. The value of the test statistic is calculated as
LO 20.5 Determine whether the medians of more than two populations differ.
We employ the Kruskal-Wallis test as the nonparametric alternative to the one-way
ANOVA F test. It is based on ranks and is used for testing the differences between the
medians of k populations. The value of the test statistic for the Kruskal-Wallis test is
where Ri and inare the rank sum and the size of the ith
LO 20.7 Determine whether two populations differ under matched-pairs sampling with
ordinal data.
We use the sign test to determine whether significant differences exist between two
matched-pairs populations of ordinal data. The value of the test statistic is computed as
is the sample proportion of positive signs. The test is valid when
n ≥ 10.
uses the old fertilizer on one half of each lot and the new 16 16
fertilizer on the other half. The follow ing table shows the 14 20
results. ⋮ ⋮
63. The table below shows a portion o f the returns for 1200 1260 1100 1600
Fidelity's Equity Income mutual fund and Vanguard's 1450 1400 1050 1325
Equity Income mutual fund from 2000 through 2010; 1150 1425 1025 1200
the full data set is on the text website, labeled Fund
a. Specify th e c o m p e tin g hypotheses to te st w h e th e r th e
Comparison.
m edian te st scores d iffe r by location.
Year Fidelity Vanguard
b. Calculate th e value o f th e te s t statistic H.
2000 3.88 13.57 c. A t th e 5% significance level, w h a t is th e critical value?
2001 − 5.02 − 2.34 d. Do th e data s u p p o rt th e psychiatrist's belief? Explain.
⋮ ⋮ ⋮
66 . An e c o n om ist w a n ts to d e te rm in e w h e th e r th e Price/
2010 15.13 14.88 Earnings (P/E) ratio is th e same fo r firm s in th re e industries.
Source : h t t p : / / f in a n c e . y a h o o . c o m . In d e p e n d e n t samples o f five firm s in each in d u s try show
th e fo llo w in g results. The data are also available on th e te x t Restaurant Inspector A's Rating Inspector B's Rating
w ebsite, labeled Industry P /E .
1 9 8
12.19
In d u s t r y A 12.44 7.28 9.96 10.51 2 5 6
14.34
in d u s t r y B 17.80 9.32 14.90 9.41 3 10 8
26.38
In d u s t r y C 24.75 16.88 16.87 16.70 4 5 4
5 2 3
a. Specify th e c o m p e tin g hypotheses to te st w h e th e r th e
m edian P/E ratios d iffe r by industry. 6 7 5
returns (in percent) for tw o of Fidelity's mutual funds: the a. Using th e sign test, specify th e co m p e tin g
Fidelity Advisor's Electronic Fund and the Fidelity Advisor's hypotheses to d e te rm in e w h e th e r th e ratings
Utilities Fund; the full data set is on the text website, labeled are sig n ifica n tly d iffe re n t b e tw e en th e tw o health
Electronics a nd Utilities. inspectors.
CASE STUDIES
1 418 167
2 491 186
⋮ ⋮ ⋮
10 885 262
S o urce: Compustat.
Data for Case Study 20.3 C o n s u m p tio n and D isposable Incom e, 1 9 8 5-2 0 0 6
Tables
z 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09
-3 .9 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000
-3 .8 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001
-3 .7 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001
-3 .6 0.0002 0.0002 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001
-3 .5 0.0002 0.0002 0.0002 0.0002 0.0002 0.0002 0.0002 0.0002 0.0002 0.0002
-3 .4 0.0003 0.0003 0.0003 0.0003 0.0003 0.0003 0.0003 0.0003 0.0003 0.0002
-3 .3 0.0005 0.0005 0.0005 0.0004 0.0004 0.0004 0.0004 0.0004 0.0004 0.0003
-3 .2 0.0007 0.0007 0.0006 0.0006 0.0006 0.0006 0.0006 0.0005 0.0005 0.0005
-3 .1 0.0010 0.0009 0.0009 0.0009 0.0008 0.0008 0.0008 0.0008 0.0007 0.0007
-3 .0 0.0013 0.0013 0.0013 0.0012 0.0012 0.0011 0.0011 0.0011 0.0010 0.0010
-2 .9 0.0019 0.0018 0.0018 0.0017 0.0016 0.0016 0.0015 0.0015 0.0014 0.0014
-2 .8 0.0026 0.0025 0.0024 0.0023 0.0023 0.0022 0.0021 0.0021 0.0020 0.0019
-2 .7 0.0035 0.0034 0.0033 0.0032 0.0031 0.0030 0.0029 0.0028 0.0027 0.0026
-2 .6 0.0047 0.0045 0.0044 0.0043 0.0041 0.0040 0.0039 0.0038 0.0037 0.0036
-2 .5 0.0062 0.0060 0.0059 0.0057 0.0055 0.0054 0.0052 0.0051 0.0049 0.0048
-2 .4 0.0082 0.0080 0.0078 0.0075 0.0073 0.0071 0.0069 0.0068 0.0066 0.0064
-2 .3 0.0107 0.0104 0.0102 0.0099 0.0096 0.0094 0.0091 0.0089 0.0087 0.0084
-2 .2 0.0139 0.0136 0.0132 0.0129 0.0125 0.0122 0.0119 0.0116 0.0113 0.0110
- 2 .1 0.0179 0.0174 0.0170 0.0166 0.0162 0.0158 0.0154 0.0150 0.0146 0.0143
-2 .0 0.0228 0.0222 0.0217 0.0212 0.0207 0.0202 0.0197 0.0192 0.0188 0.0183
-1 .9 0.0287 0.0281 0.0274 0.0268 0.0262 0.0256 0.0250 0.0244 0.0239 0.0233
-1 .8 0.0359 0.0351 0.0344 0.0336 0.0329 0.0322 0.0314 0.0307 0.0301 0.0294
-1 .7 0.0446 0.0436 0.0427 0.0418 0.0409 0.0401 0.0392 0.0384 0.0375 0.0367
-1 .6 0.0548 0.0537 0.0526 0.0516 0.0505 0.0495 0.0485 0.0475 0.0465 0.0455
-1 .5 0.0668 0.0655 0.0643 0.0630 0.0618 0.0606 0.0594 0.0582 0.0571 0.0559
-1 .4 0.0808 0.0793 0.0778 0.0764 0.0749 0.0735 0.0721 0.0708 0.0694 0.0681
-1 .3 0.0968 0.0951 0.0934 0.0918 0.0901 0.0885 0.0869 0.0853 0.0838 0.0823
-1 .2 0.1151 0.1131 0.1112 0.1093 0.1075 0.1056 0.1038 0.1020 0.1003 0.0985
- 1.1 0.1357 0.1335 0.1314 0.1292 0.1271 0.1251 0.1230 0.1210 0.1190 0.1170
-1 .0 0.1587 0.1562 0.1539 0.1515 0.1492 0.1469 0.1446 0.1423 0.1401 0.1379
-0 .9 0.1841 0.1814 0.1788 0.1762 0.1736 0.1711 0.1685 0.1660 0.1635 0.1611
-0 .8 0.2119 0.2090 0.2061 0.2033 0.2005 0.1977 0.1949 0.1922 0.1894 0.1867
-0 .7 0.2420 0.2389 0.2358 0.2327 0.2296 0.2266 0.2236 0.2206 0.2177 0.2148
-0 .6 0.2743 0.2709 0.2676 0.2643 0.2611 0.2578 0.2546 0.2514 0.2483 0.2451
-0 .5 0.3085 0.3050 0.3015 0.2981 0.2946 0.2912 0.2877 0.2843 0.2810 0.2776
-0 .4 0.3446 0.3409 0.3372 0.3336 0.3300 0.3264 0.3228 0.3192 0.3156 0.3121
-0 .3 0.3821 0.3783 0.3745 0.3707 0.3669 0.3632 0.3594 0.3557 0.3520 0.3483
-0 .2 0.4207 0.4168 0.4129 0.4090 0.4052 0.4013 0.3974 0.3936 0.3897 0.3859
-0 .1 0.4602 0.4562 0.4522 0.4483 0.4443 0.4404 0.4364 0.4325 0.4286 0.4247
-0 .0 0.5000 0.4960 0.4920 0.4880 0.4840 0.4801 0.4761 0.4721 0.4681 0.4641
z 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09
0.0 0.5000 0.5040 0.5080 0.5120 0.5160 0.5199 0.5239 0.5279 0.5319 0.5359
0.1 0.5398 0.5438 0.5478 0.5517 0.5557 0.5596 0.5636 0.5675 0.5714 0.5753
0.2 0.5793 0.5832 0.5871 0.5910 0.5948 0.5987 0.6026 0.6064 0.6103 0.6141
0.3 0.6179 0.6217 0.6255 0.6293 0.6331 0.6368 0.6406 0.6443 0.6480 0.6517
0.4 0.6554 0.6591 0.6628 0.6664 0.6700 0.6736 0.6772 0.6808 0.6844 0.6879
0.5 0.6915 0.6950 0.6985 0.7019 0.7054 0.7088 0.7123 0.7157 0.7190 0.7224
0.6 0.7257 0.7291 0.7324 0.7357 0.7389 0.7422 0.7454 0.7486 0.7517 0.7549
0.7 0.7580 0.7611 0.7642 0.7673 0.7704 0.7734 0.7764 0.7794 0.7823 0.7852
0.8 0.7881 0.7910 0.7939 0.7967 0.7995 0.8023 0.8051 0.8078 0.8106 0.8133
0.9 0.8159 0.8186 0.8212 0.8238 0.8264 0.8289 0.8315 0.8340 0.8365 0.8389
1.0 0.8413 0.8438 0.8461 0.8485 0.8508 0.8531 0.8554 0.8577 0.8599 0.8621
1.1 0.8643 0.8665 0.8686 0.8708 0.8729 0.8749 0.8770 0.8790 0.8810 0.8830
1.2 0.8849 0.8869 0.8888 0.8907 0.8925 0.8944 0.8962 0.8980 0.8997 0.9015
1.3 0.9032 0.9049 0.9066 0.9082 0.9099 0.9115 0.9131 0.9147 0.9162 0.9177
1.4 0.9192 0.9207 0.9222 0.9236 0.9251 0.9265 0.9279 0.9292 0.9306 0.9319
1.5 0.9332 0.9345 0.9357 0.9370 0.9382 0.9394 0.9406 0.9418 0.9429 0.9441
1.6 0.9452 0.9463 0.9474 0.9484 0.9495 0.9505 0.9515 0.9525 0.9535 0.9545
1.7 0.9554 0.9564 0.9573 0.9582 0.9591 0.9599 0.9608 0.9616 0.9625 0.9633
1.8 0.9641 0.9649 0.9656 0.9664 0.9671 0.9678 0.9686 0.9693 0.9699 0.9706
1.9 0.9713 0.9719 0.9726 0.9732 0.9738 0.9744 0.9750 0.9756 0.9761 0.9767
2.0 0.9772 0.9778 0.9783 0.9788 0.9793 0.9798 0.9803 0.9808 0.9812 0.9817
2.1 0.9821 0.9826 0.9830 0.9834 0.9838 0.9842 0.9846 0.9850 0.9854 0.9857
2.2 0.9861 0.9864 0.9868 0.9871 0.9875 0.9878 0.9881 0.9884 0.9887 0.9890
2.3 0.9893 0.9896 0.9898 0.9901 0.9904 0.9906 0.9909 0.9911 0.9913 0.9916
2.4 0.9918 0.9920 0.9922 0.9925 0.9927 0.9929 0.9931 0.9932 0.9934 0.9936
2.5 0.9938 0.9940 0.9941 0.9943 0.9945 0.9946 0.9948 0.9949 0.9951 0.9952
2.6 0.9953 0.9955 0.9956 0.9957 0.9959 0.9960 0.9961 0.9962 0.9963 0.9964
2.7 0.9965 0.9966 0.9967 0.9968 0.9969 0.9970 0.9971 0.9972 0.9973 0.9974
2.8 0.9974 0.9975 0.9976 0.9977 0.9977 0.9978 0.9979 0.9979 0.9980 0.9981
2.9 0.9981 0.9982 0.9982 0.9983 0.9984 0.9984 0.9985 0.9985 0.9986 0.9986
3.0 0.9987 0.9987 0.9987 0.9988 0.9988 0.9989 0.9989 0.9989 0.9990 0.9990
3.1 0.9990 0.9991 0.9991 0.9991 0.9992 0.9992 0.9992 0.9992 0.9993 0.9993
3.2 0.9993 0.9993 0.9994 0.9994 0.9994 0.9994 0.9994 0.9995 0.9995 0.9995
3.3 0.9995 0.9995 0.9995 0.9996 0.9996 0.9996 0.9996 0.9996 0.9996 0.9997
3.4 0.9997 0.9997 0.9997 0.9997 0.9997 0.9997 0.9997 0.9997 0.9997 0.9998
3.5 0.9998 0.9998 0.9998 0.9998 0.9998 0.9998 0.9998 0.9998 0.9998 0.9998
3.6 0.9998 0.9998 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999
3.7 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999
3.8 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999
3.9 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999
α
df 0.20 0.10 0.05 0.025 0.01 0.005
α
df 0.20 0.10 0.05 0.025 0.01 0.005
31 0.853 1.309 1.696 2.040 2.453 2.744
32 0.853 1.309 1.694 2.037 2.449 2.738
33 0.853 1.308 1.692 2.035 2.445 2.733
34 0.852 1.307 1.691 2.032 2.441 2.728
35 0.852 1.306 1.690 2.030 2.438 2.724
36 0.852 1.306 1.688 2.028 2.434 2.719
37 0.851 1.305 1.687 2.026 2.431 2.715
38 0.851 1.304 1.686 2.024 2.429 2.712
39 0.851 1.304 1.685 2.023 2.426 2.708
40 0.851 1.303 1.684 2.021 2.423 2.704
α
df 0.995 0.990 0.975 0.950 0.900 0.100 0.050 0.025 0.010 0.005
1 0.000 0.000 0.001 0.004 0.016 2.706 3.841 5.024 6.635 7.879
2 0.010 0.020 0.051 0.103 0.211 4.605 5.991 7.378 9.210 10.597
3 0.072 0.115 0.216 0.352 0.584 6.251 7.815 9.348 11.345 12.838
4 0.207 0.297 0.484 0.711 1.064 7.779 9.488 11.143 13.277 14.860
5 0.412 0.554 0.831 1.145 1.610 9.236 11.070 12.833 15.086 16.750
6 0.676 0.872 1.237 1.635 2.204 10.645 12.592 14.449 16.812 18.548
7 0.989 1.239 1.690 2.167 2.833 12.017 14.067 16.013 18.475 20.278
8 1.344 1.646 2.180 2.733 3.490 13.362 15.507 17.535 20.090 21.955
9 1.735 2.088 2.700 3.325 4.168 14.684 16.919 19.023 21.666 23.589
10 2.156 2.558 3.247 3.940 4.865 15.987 18.307 20.483 23.209 25.188
11 2.603 3.053 3.816 4.575 5.578 17.275 19.675 21.920 24.725 26.757
12 3.074 3.571 4.404 5.226 6.304 18.549 21.026 23.337 26.217 28.300
13 3.565 4.107 5.009 5.892 7.042 19.812 22.362 24.736 27.688 29.819
14 4.075 4.660 5.629 6.571 7.790 21.064 23.685 26.119 29.141 31.319
15 4.601 5.229 6.262 7.261 8.547 22.307 24.996 27.488 30.578 32.801
16 5.142 5.812 6.908 7.962 9.312 23.542 26.296 28.845 32.000 34.267
17 5.697 6.408 7.564 8.672 10.085 24.769 27.587 30.191 33.409 35.718
18 6.265 7.015 8.231 9.390 10.865 25.989 28.869 31.526 34.805 37.156
19 6.844 7.633 8.907 10.117 11.651 27.204 30.144 32.852 36.191 38.582
20 7.434 8.260 9.591 10.851 12.443 28.412 31.410 34.170 37.566 39.997
21 8.034 8.897 10.283 11.591 13.240 29.615 32.671 35.479 38.932 41.401
22 8.643 9.542 10.982 12.338 14.041 30.813 33.924 36.781 40.289 42.796
23 9.260 10.196 11.689 13.091 14.848 32.007 35.172 38.076 41.638 44.181
24 9.886 10.856 12.401 13.848 15.659 33.196 36.415 39.364 42.980 45.559
25 10.520 11.524 13.120 14.611 16.473 34.382 37.652 40.646 44.314 46.928
26 11.160 12.198 13.844 15.379 17.292 35.563 38.885 41.923 45.642 48.290
27 11.808 12.879 14.573 16.151 18.114 36.741 40.113 43.195 46.963 49.645
28 12.461 13.565 15.308 16.928 18.939 37.916 41.337 44.461 48.278 50.993
29 13.121 14.256 16.047 17.708 19.768 39.087 42.557 45.722 49.588 52.336
30 13.787 14.953 16.791 18.493 20.599 40.256 43.773 46.979 50.892 53.672
TABLE 3 (Continued)
df 0.995 0.990 0.975 0.950 0.900 0.100 0.050 0.025 0.010 0.005
31 14.458 15.655 17.539 19.281 21.434 41.422 44.985 48.232 52.191 55.003
32 15.134 16.362 18.291 20.072 22.271 42.585 46.194 49.480 53.486 56.328
33 15.815 17.074 19.047 20.867 23.110 43.745 47.400 50.725 54.776 57.648
34 16.501 17.789 19.806 21.664 23.952 44.903 48.602 51.966 56.061 58.964
35 17.192 18.509 20.569 22.465 24.797 46.059 49.802 53.203 57.342 60.275
36 17.887 19.233 21.336 23.269 25.643 47.212 50.998 54.437 58.619 61.581
37 18.586 19.960 22.106 24.075 26.492 48.363 52.192 55.668 59.893 62.883
38 19.289 20.691 22.878 24.884 27.343 49.513 53.384 56.896 61.162 64.181
39 19.996 21.426 23.654 25.695 28.196 50.660 54.572 58.120 62.428 65.476
40 20.707 22.164 24.433 26.509 29.051 51.805 55.758 59.342 63.691 66.766
41 21.421 22.906 25.215 27.326 29.907 52.949 56.942 60.561 64.950 68.053
42 22.138 23.650 25.999 28.144 30.765 54.090 58.124 61.777 66.206 69.336
43 22.859 24.398 26.785 28.965 31.625 55.230 59.304 62.990 67.459 70.616
44 23.584 25.148 27.575 29.787 32.487 56.369 60.481 64.201 68.710 71.893
45 24.311 25.901 28.366 30.612 33.350 57.505 61.656 65.410 69.957 73.166
46 25.041 26.657 29.160 31.439 34.215 58.641 62.830 66.617 71.201 74.437
47 25.775 27.416 29.956 32.268 35.081 59.774 64.001 67.821 72.443 75.704
48 26.511 28.177 30.755 33.098 35.949 60.907 65.171 69.023 73.683 76.969
49 27.249 28.941 31.555 33.930 36.818 62.038 66.339 70.222 74.919 78.231
50 27.991 29.707 32.357 34.764 37.689 63.167 67.505 71.420 76.154 79.490
55 31.735 33.570 36.398 38.958 42.060 68.796 73.311 77.380 82.292 85.749
60 35.534 37.485 40.482 43.188 46.459 74.397 79.082 83.298 88.379 91.952
65 39.383 41.444 44.603 47.450 50.883 79.973 84.821 89.177 94.422 98.105
70 43.275 45.442 48.758 51.739 55.329 85.527 90.531 95.023 100.425 104.215
75 47.206 49.475 52.942 56.054 59.795 91.061 96.217 100.839 106.393 110.286
80 51.172 53.540 57.153 60.391 64.278 96.578 101.879 106.629 112.329 116.321
85 55.170 57.634 61.389 64.749 68.777 102.079 107.522 112.393 118.236 122.325
90 59.196 61.754 65.647 69.126 73.291 107.565 113.145 118.136 124.116 128.299
95 63.250 65.898 69.925 73.520 77.818 113.038 118.752 123.858 129.973 134.247
100 67.328 70.065 74.222 77.929 82.358 118.498 124.342 129.561 135.807 140.169
d f1
df2 α 1 2 3 4 5 6 7 8 9 10 15 25 50 100 500
1 0.10 39.86 49.5 53.59 55.83 57.24 58.2 58.91 59.44 59.86 60.19 61.22 62.05 62.69 63.01 63.26
0.05 161.45 199.50 215.71 224.58 230.16 233.99 236.77 238.88 240.54 241.88 245.95 249.26 251.77 253.04 254.06
0.025 647.79 799.50 864.16 899.58 921.85 937.11 948.22 956.66 963.28 968.63 984.87 998.08 1008.12 1013.17 1017.24
0.01 4052.18 4999.50 5403.35 5624.58 5763.65 5858.99 5928.36 5981.07 6022.47 6055.85 6157.28 6239.83 6302.52 6334.11 6359.50
2 0.10 8.53 9.00 9.16 9.24 9.29 9.33 9.35 9.37 9.38 9.39 9.42 9.45 9.47 9.48 9.49
0.05 18.51 19.00 19.16 19.25 19.30 19.33 19.35 19.37 19.38 19.40 19.43 19.46 19.48 19.49 19.49
0.025 38.51 39.00 39.17 39.25 39.30 39.33 39.36 39.37 39.39 39.40 39.43 39.46 39.48 39.49 39.50
0.01 98.50 99.00 99.17 99.25 99.30 99.33 99.36 99.37 99.39 99.40 99.43 99.46 99.48 99.49 99.50
3 0.10 5.54 5.46 5.39 5.34 5.31 5.28 5.27 5.25 5.24 5.23 5.20 5.17 5.15 5.14 5.14
0.05 10.13 9.55 9.28 9.12 9.01 8.94 8.89 8.85 8.81 8.79 8.70 8.63 8.58 8.55 8.53
0.025 17.44 16.04 15.44 15.10 14.88 14.73 14.62 14.54 14.47 14.42 14.25 14.12 14.01 13.96 13.91
0.01 34.12 30.82 29.46 28.71 28.24 27.91 27.67 27.49 27.35 27.23 26.87 26.58 26.35 26.24 26.15
4 0.10 4.54 4.32 4.19 4.11 4.05 4.01 3.98 3.95 3.94 3.92 3.87 3.83 3.80 3.78 3.76
0.05 7.71 6.94 6.59 6.39 6.26 6.16 6.09 6.04 6.00 5.96 5.86 5.77 5.70 5.66 5.64
0.025 12.22 10.65 9.98 9.60 9.36 9.20 9.07 8.98 8.90 8.84 8.66 8.5 8.38 8.32 8.27
0.01 21.20 18.00 16.69 15.98 15.52 15.21 14.98 14.80 14.66 14.55 14.20 13.91 13.69 13.58 13.49
5 0.10 4.06 3.78 3.62 3.52 3.45 3.4 3.37 3.34 3.32 3.30 3.24 3.19 3.15 3.13 3.11
0.05 6.61 5.79 5.41 5.19 5.05 4.95 4.88 4.82 4.77 4.74 4.62 4.52 4.44 4.41 4.37
0.025 10.01 8.43 7.76 7.39 7.15 6.98 6.85 6.76 6.68 6.62 6.43 6.27 6.14 6.08 6.03
0.01 16.26 13.27 12.06 11.39 10.97 10.67 10.46 10.29 10.16 10.05 9.72 9.45 9.24 9.13 9.04
6 0.10 3.78 3.46 3.29 3.18 3.11 3.05 3.01 2.98 2.96 2.94 2.87 2.81 2.77 2.75 2.73
0.05 5.99 5.14 4.76 4.53 4.39 4.28 4.21 4.15 4.10 4.06 3.94 3.83 3.75 3.71 3.68
0.025 8.81 7.26 6.60 6.23 5.99 5.82 5.70 5.6 5.52 5.46 5.27 5.11 4.98 4.92 4.86
0.01 13.75 10.92 9.78 9.15 8.75 8.47 8.26 8.10 7.98 7.87 7.56 7.30 7.09 6.99 6.90
7 0.10 3.59 3.26 3.07 2.96 2.88 2.83 2.78 2.75 2.72 2.70 2.63 2.57 2.52 2.50 2.48
0.05 5.59 4.74 4.35 4.12 3.97 3.87 3.79 3.73 3.68 3.64 3.51 3.4 3.32 3.27 3.24
0.025 8.07 6.54 5.89 5.52 5.29 5.12 4.99 4.90 4.82 4.76 4.57 4.4 4.28 4.21 4.16
0.01 12.25 9.55 8.45 7.85 7.46 7.19 6.99 6.84 6.72 6.62 6.31 6.06 5.86 5.75 5.67
d f1
d f2 α 1 2 3 4 5 6 7 8 9 10 15 25 50 100 500
8 0.10 3.46 3.11 2.92 2.81 2.73 2.67 2.62 2.59 2.56 2.54 2.46 2.40 2.35 2.32 2.30
0.05 5.32 4.46 4.07 3.84 3.69 3.58 3.50 3.44 3.39 3.35 3.22 3.11 3.02 2.97 2.94
0.025 7.57 6.06 5.42 5.05 4.82 4.65 4.53 4.43 4.36 4.30 4.1 3.94 3.81 3.74 3.68
0.01 11.26 8.65 7.59 7.01 6.63 6.37 6.18 6.03 5.91 5.81 5.52 5.26 5.07 4.96 4.88
9 0.10 3.36 3.01 2.81 2.69 2.61 2.55 2.51 2.47 2.44 2.42 2.34 2.27 2.22 2.19 2.17
0.05 5.12 4.26 3.86 3.63 3.48 3.37 3.29 3.23 3.18 3.14 3.01 2.89 2.80 2.76 2.72
0.025 7.21 5.71 5.08 4.72 4.48 4.32 4.20 4.10 4.03 3.96 3.77 3.6 3.47 3.40 3.35
0.01 10.56 8.02 6.99 6.42 6.06 5.8 5.61 5.47 5.35 5.26 4.96 4.71 4.52 4.41 4.33
10 0.10 3.29 2.92 2.73 2.61 2.52 2.46 2.41 2.38 2.35 2.32 2.24 2.17 2.12 2.09 2.06
0.05 4.96 4.1 3.71 3.48 3.33 3.22 3.14 3.07 3.02 2.98 2.85 2.73 2.64 2.59 2.55
0.025 6.94 5.46 4.83 4.47 4.24 4.07 3.95 3.85 3.78 3.72 3.52 3.35 3.22 3.15 3.09
0.01 10.04 7.56 6.55 5.99 5.64 5.39 5.20 5.06 4.94 4.85 4.56 4.31 4.12 4.01 3.93
11 0.10 3.23 2.86 2.66 2.54 2.45 2.39 2.34 2.30 2.27 2.25 2.17 2.10 2.04 2.01 1.98
0.05 4.84 3.98 3.59 3.36 3.20 3.09 3.01 2.95 2.90 2.85 2.72 2.60 2.51 2.46 2.42
0.025 6.72 5.26 4.63 4.28 4.04 3.88 3.76 3.66 3.59 3.53 3.33 3.16 3.03 2.96 2.90
0.01 9.65 7.21 6.22 5.67 5.32 5.07 4.89 4.74 4.63 4.54 4.25 4.01 3.81 3.71 3.62
12 0.10 3.18 2.81 2.61 2.48 2.39 2.33 2.28 2.24 2.21 2.19 2.10 2.03 1.97 1.94 1.91
0.05 4.75 3.89 3.49 3.26 3.11 3.00 2.91 2.85 2.80 2.75 2.62 2.50 2.40 2.35 2.31
0.025 6.55 5.10 4.47 4.12 3.89 3.73 3.61 3.51 3.44 3.37 3.18 3.01 2.87 2.80 2.74
0.01 9.33 6.93 5.95 5.41 5.06 4.82 4.64 4.50 4.39 4.30 4.01 3.76 3.57 3.47 3.38
13 0.10 3.14 2.76 2.56 2.43 2.35 2.28 2.23 2.20 2.16 2.14 2.05 1.98 1.92 1.88 1.85
0.05 4.67 3.81 3.41 3.18 3.03 2.92 2.83 2.77 2.71 2.67 2.53 2.41 2.31 2.26 2.22
0.025 6.41 4.97 4.35 4.00 3.77 3.60 3.48 3.39 3.31 3.25 3.05 2.88 2.74 2.67 2.61
0.01 9.07 6.70 5.74 5.21 4.86 4.62 4.44 4.30 4.19 4.10 3.82 3.57 3.38 3.27 3.19
14 0.10 3.10 2.73 2.52 2.39 2.31 2.24 2.19 2.15 2.12 2.10 2.01 1.93 1.87 1.83 1.80
0.05 4.60 3.74 3.34 3.11 2.96 2.85 2.76 2.70 2.65 2.60 2.46 2.34 2.24 2.19 2.14
0.025 6.30 4.86 4.24 3.89 3.66 3.50 3.38 3.29 3.21 3.15 2.95 2.78 2.64 2.56 2.50
0.01 8.86 6.51 5.56 5.04 4.69 4.46 4.28 4.14 4.03 3.94 3.66 3.41 3.22 3.11 3.03
15 0.10 3.07 2.7 2.49 2.36 2.27 2.21 2.16 2.12 2.09 2.06 1.97 1.89 1.83 1.79 1.76
0.05 4.54 3.68 3.29 3.06 2.90 2.79 2.71 2.64 2.59 2.54 2.40 2.28 2.18 2.12 2.08
0.025 6.20 4.77 4.15 3.80 3.58 3.41 3.29 3.20 3.12 3.06 2.86 2.69 2.55 2.47 2.41
0.01 8.68 6.36 5.42 4.89 4.56 4.32 4.14 4.00 3.89 3.80 3.52 3.28 3.08 2.98 2.89
16 0.10 3.05 2.67 2.46 2.33 2.24 2.18 2.13 2.09 2.06 2.03 1.94 1.86 1.79 1.76 1.73
0.05 4.49 3.63 3.24 3.01 2.85 2.74 2.66 2.59 2.54 2.49 2.35 2.23 2.12 2.07 2.02
0.025 6.12 4.69 4.08 3.73 3.50 3.34 3.22 3.12 3.05 2.99 2.79 2.61 2.47 2.40 2.33
0.01 8.53 6.23 5.29 4.77 4.44 4.20 4.03 3.89 3.78 3.69 3.41 3.16 2.97 2.86 2.78
TABLE 4 (Continued)
df1
df2 α 1 2 3 4 5 6 7 8 9 10 15 25 50 100 500
17 0.10 3.03 2.64 2.44 2.31 2.22 2.15 2.10 2.06 2.03 2.00 1.91 1.83 1.76 1.73 1.69
0.05 4.45 3.59 3.20 2.96 2.81 2.70 2.61 2.55 2.49 2.45 2.31 2.18 2.08 2.02 1.97
0.025 6.04 4.62 4.01 3.66 3.44 3.28 3.16 3.06 2.98 2.92 2.72 2.55 2.41 2.33 2.26
0.01 8.40 6.11 5.18 4.67 4.34 4.10 3.93 3.79 3.68 3.59 3.31 3.07 2.87 2.76 2.68
18 0.10 3.01 2.62 2.42 2.29 2.20 2.13 2.08 2.04 2.00 1.98 1.89 1.80 1.74 1.70 1.67
0.05 4.41 3.55 3.16 2.93 2.77 2.66 2.58 2.51 2.46 2.41 2.27 2.14 2.04 1.98 1.93
0.025 5.98 4.56 3.95 3.61 3.38 3.22 3.10 3.01 2.93 2.87 2.67 2.49 2.35 2.27 2.20
0.01 8.29 6.01 5.09 4.58 4.25 4.01 3.84 3.71 3.60 3.51 3.23 2.98 2.78 2.68 2.59
19 0.10 2.99 2.61 2.40 2.27 2.18 2.11 2.06 2.02 1.98 1.96 1.86 1.78 1.71 1.67 1.64
0.05 4.38 3.52 3.13 2.90 2.74 2.63 2.54 2.48 2.42 2.38 2.23 2.11 2.00 1.94 1.89
0.025 5.92 4.51 3.90 3.56 3.33 3.17 3.05 2.96 2.88 2.82 2.62 2.44 2.30 2.22 2.15
0.01 8.18 5.93 5.01 4.50 4.17 3.94 3.77 3.63 3.52 3.43 3.15 2.91 2.71 2.60 2.51
20 0.10 2.97 2.59 2.38 2.25 2.16 2.09 2.04 2.00 1.96 1.94 1.84 1.76 1.69 1.65 1.62
0.05 4.35 3.49 3.10 2.87 2.71 2.60 2.51 2.45 2.39 2.35 2.20 2.07 1.97 1.91 1.86
0.025 5.87 4.46 3.86 3.51 3.29 3.13 3.01 2.91 2.84 2.77 2.57 2.40 2.25 2.17 2.10
0.01 8.10 5.85 4.94 4.43 4.10 3.87 3.70 3.56 3.46 3.37 3.09 2.84 2.64 2.54 2.44
21 0.10 2.96 2.57 2.36 2.23 2.14 2.08 2.02 1.98 1.95 1.92 1.83 1.74 1.67 1.63 1.60
0.05 4.32 3.47 3.07 2.84 2.68 2.57 2.49 2.42 2.37 2.32 2.18 2.05 1.94 1.88 1.83
0.025 5.83 4.42 3.82 3.48 3.25 3.09 2.97 2.87 2.80 2.73 2.53 2.36 2.21 2.13 2.06
0.01 8.02 5.78 4.87 4.37 4.04 3.81 3.64 3.51 3.40 3.31 3.03 2.79 2.58 2.48 2.38
22 0.10 2.95 2.56 2.35 2.22 2.13 2.06 2.01 1.97 1.93 1.90 1.81 1.73 1.65 1.61 1.58
0.05 4.30 3.44 3.05 2.82 2.66 2.55 2.46 2.40 2.34 2.30 2.15 2.02 1.91 1.85 1.80
0.025 5.79 4.38 3.78 3.44 3.22 3.05 2.93 2.84 2.76 2.70 2.50 2.32 2.17 2.09 2.02
0.01 7.95 5.72 4.82 4.31 3.99 3.76 3.59 3.45 3.35 3.26 2.98 2.73 2.53 2.42 2.33
23 0.10 2.94 2.55 2.34 2.21 2.11 2.05 1.99 1.95 1.92 1.89 1.80 1.71 1.64 1.59 1.56
0.05 4.28 3.42 3.03 2.80 2.64 2.53 2.44 2.37 2.32 2.27 2.13 2.00 1.88 1.82 1.77
0.025 5.75 4.35 3.75 3.41 3.18 3.02 2.90 2.81 2.73 2.67 2.47 2.29 2.14 2.06 1.99
0.01 7.88 5.66 4.76 4.26 3.94 3.71 3.54 3.41 3.30 3.21 2.93 2.69 2.48 2.37 2.28
24 0.10 2.93 2.54 2.33 2.19 2.10 2.04 1.98 1.94 1.91 1.88 1.78 1.70 1.62 1.58 1.54
0.05 4.26 3.40 3.01 2.78 2.62 2.51 2.42 2.36 2.30 2.25 2.11 1.97 1.86 1.80 1.75
0.025 5.72 4.32 3.72 3.38 3.15 2.99 2.87 2.78 2.70 2.64 2.44 2.26 2.11 2.02 1.95
0.01 7.82 5.61 4.72 4.22 3.90 3.67 3.50 3.36 3.26 3.17 2.89 2.64 2.44 2.33 2.24
df1
df2 α 1 2 3 4 5 6 7 8 9 10 15 25 50 100 500
25 0.10 2.92 2.53 2.32 2.18 2.09 2.02 1.97 1.93 1.89 1.87 1.77 1.68 1.61 1.56 1.53
0.05 4.24 3.39 2.99 2.76 2.60 2.49 2.40 2.34 2.28 2.24 2.09 1.96 1.84 1.78 1.73
0.025 5.69 4.29 3.69 3.35 3.13 2.97 2.85 2.75 2.68 2.61 2.41 2.23 2.08 2.00 1.92
0.01 7.77 5.57 4.68 4.18 3.85 3.63 3.46 3.32 3.22 3.13 2.85 2.60 2.40 2.29 2.19
26 0.10 2.91 2.52 2.31 2.17 2.08 2.01 1.96 1.92 1.88 1.86 1.76 1.67 1.59 1.55 1.51
0.05 4.23 3.37 2.98 2.74 2.59 2.47 2.39 2.32 2.27 2.22 2.07 1.94 1.82 1.76 1.71
0.025 5.66 4.27 3.67 3.33 3.10 2.94 2.82 2.73 2.65 2.59 2.39 2.21 2.05 1.97 1.90
0.01 7.72 5.53 4.64 4.14 3.82 3.59 3.42 3.29 3.18 3.09 2.81 2.57 2.36 2.25 2.16
27 0.10 2.90 2.51 2.30 2.17 2.07 2.00 1.95 1.91 1.87 1.85 1.75 1.66 1.58 1.54 1.50
0.05 4.21 3.35 2.96 2.73 2.57 2.46 2.37 2.31 2.25 2.20 2.06 1.92 1.81 1.74 1.69
0.025 5.63 4.24 3.65 3.31 3.08 2.92 2.80 2.71 2.63 2.57 2.36 2.18 2.03 1.94 1.87
0.01 7.68 5.49 4.60 4.11 3.78 3.56 3.39 3.26 3.15 3.06 2.78 2.54 2.33 2.22 2.12
28 0.10 2.89 2.50 2.29 2.16 2.06 2.00 1.94 1.90 1.87 1.84 1.74 1.65 1.57 1.53 1.49
0.05 4.20 3.34 2.95 2.71 2.56 2.45 2.36 2.29 2.24 2.19 2.04 1.91 1.79 1.73 1.67
0.025 5.61 4.22 3.63 3.29 3.06 2.90 2.78 2.69 2.61 2.55 2.34 2.16 2.01 1.92 1.85
0.01 7.64 5.45 4.57 4.07 3.75 3.53 3.36 3.23 3.12 3.03 2.75 2.51 2.30 2.19 2.09
29 0.10 2.89 2.50 2.28 2.15 2.06 1.99 1.93 1.89 1.86 1.83 1.73 1.64 1.56 1.52 1.48
0.05 4.18 3.33 2.93 2.70 2.55 2.43 2.35 2.28 2.22 2.18 2.03 1.89 1.77 1.71 1.65
0.025 5.59 4.20 3.61 3.27 3.04 2.88 2.76 2.67 2.59 2.53 2.32 2.14 1.99 1.90 1.83
0.01 7.60 5.42 4.54 4.04 3.73 3.50 3.33 3.20 3.09 3.00 2.73 2.48 2.27 2.16 2.06
30 0.10 2.88 2.49 2.28 2.14 2.05 1.98 1.93 1.88 1.85 1.82 1.72 1.63 1.55 1.51 1.47
0.05 4.17 3.32 2.92 2.69 2.53 2.42 2.33 2.27 2.21 2.16 2.01 1.88 1.76 1.70 1.64
0.025 5.57 4.18 3.59 3.25 3.03 2.87 2.75 2.65 2.57 2.51 2.31 2.12 1.97 1.88 1.81
0.01 7.56 5.39 4.51 4.02 3.70 3.47 •3.30 3.17 3.07 2.98 2.70 2.45 2.25 2.13 2.03
50 0.10 2.81 2.41 2.20 2.06 1.97 1.90 1.84 1.80 1.76 1.73 1.63 1.53 1.44 1.39 1.34
0.05 4.03 3.18 2.79 2.56 2.40 2.29 2.20 2.13 2.07 2.03 1.87 1.73 1.60 1.52 1.46
0.025 5.34 3.97 3.39 3.05 2.83 2.67 2.55 2.46 2.38 2.32 2.11 1.92 1.75 1.66 1.57
0.01 7.17 5.06 4.20 3.72 3.41 3.19 3.02 2.89 2.78 2.70 2.42 2.17 1.95 1.82 1.71
100 0.10 2.76 2.36 2.14 2.00 1.91 1.83 1.78 1.73 1.69 1.66 1.56 1.45 1.35 1.29 1.23
0.05 3.94 3.09 2.70 2.46 2.31 2.19 2.10 2.03 1.97 1.93 1.77 1.62 1.48 1.39 1.31
0.025 5.18 3.83 3.25 2.92 2.70 2.54 2.42 2.32 2.24 2.18 1.97 1.77 1.59 1.48 1.38
0.01 6.90 4.82 3.98 3.51 3.21 2.99 2.82 2.69 2.59 2.50 2.22 1.97 1.74 1.60 1.47
500 0.10 2.72 2.31 2.09 1.96 1.86 1.79 1.73 1.68 1.64 1.61 1.5 1.39 1.28 1.21 1.12
0.05 3.86 3.01 2.62 2.39 2.23 2.12 2.03 1.96 1.9 1.85 1.69 1.53 1.38 1.28 1.16
0.025 5.05 3.72 3.14 2.81 2.59 2.43 2.31 2.22 2.14 2.07 1.86 1.65 1.46 1.34 1.19
0.01 6.69 4.65 3.82 3.36 3.05 2.84 2.68 2.55 2.44 2.36 2.07 1.81 1.57 1.41 1.23
0.05 3.64 4.60 5.22 5.67 6.03 6.33 6.58 6.80 6.99 7.17 7.32
5
0.01 5.70 6.98 7.80 8.42 8.91 9.32 9.67 9.97 10.24 10.48 10.70
0.05 3.46 4.34 4.90 5.30 5.63 5.90 6.12 6.32 6.49 6.65 6.79
6
0.01 5.24 6.33 7.03 7.56 7.97 8.32 8.61 8.87 9.10 9.30 9.48
0.05 3.34 4.16 4.68 5.06 5.36 5.61 5.82 6.00 6.16 6.30 6.43
7
0.01 4.95 5.92 6.54 7.01 7.37 7.68 7.94 8.17 8.37 8.55 8.71
0.05 3.26 4.04 4.53 4.89 5.17 5.40 5.60 5.77 5.92 6.05 6.18
8
0.01 4.75 5.64 6.20 6.62 6.96 7.24 7.47 7.68 7.86 8.03 8.18
0.05 3.20 3.95 4.41 4.76 5.02 5.24 5.43 5.59 5.74 5.87 5.98
9
0.01 4.60 5.43 5.96 6.35 6.66 6.91 7.13 7.33 7.49 7.65 7.78
0.05 3.15 3.88 4.33 4.65 4.91 5.12 5.30 5.46 5.60 5.72 5.83
10
0.01 4.48 5.27 5.77 6.14 6.43 6.67 6.87 7.05 7.21 7.36 7.49
0.05 3.11 3.82 4.26 4.57 4.82 5.03 5.20 5.35 5.49 5.61 5.71
11
0.01 4.39 5.15 5.62 5.97 6.25 6.48 6.67 6.84 6.99 7.13 7.25
0.05 3.08 3.77 4.20 4.51 4.75 4.95 5.12 5.27 5.39 5.51 5.61
12
0.01 4.32 5.05 5.50 5.84 6.10 6.32 6.51 6.67 6.81 6.94 7.06
0.05 3.06 3.73 4.15 4.45 4.69 4.88 5.05 5.19 5.32 5.43 5.53
13
0.01 4.26 4.96 5.40 5.73 5.98 6.19 6.37 6.53 6.67 6.79 6.90
0.05 3.03 3.70 4.11 4.41 4.64 4.83 4.99 5.13 5.25 5.36 5.46
14
0.01 4.21 4.89 5.32 5.63 5.88 6.08 6.26 6.41 6.54 6.66 6.77
0.05 3.01 3.67 4.08 4.37 4.59 4.78 4.94 5.08 5.20 5.31 5.40
15
0.01 4.17 4.84 5.25 5.56 5.80 5.99 6.16 6.31 6.44 6.55 6.66
0.05 3.00 3.65 4.05 4.33 4.56 4.74 4.90 5.03 5.15 5.26 5.35
16
0.01 4.13 4.79 5.19 5.49 5.72 5.92 6.08 6.22 6.35 6.46 6.56
0.05 2.98 3.63 4.02 4.30 4.52 4.70 4.86 4.99 5.11 5.21 5.31
17
0.01 4.10 4.74 5.14 5.43 5.66 5.85 6.01 6.15 6.27 6.38 6.48
0.05 2.97 3.61 4.00 4.28 4.49 4.67 4.82 4.96 5.07 5.17 5.27
18
0.01 4.07 4.70 5.09 5.38 5.60 5.79 5.94 6.08 6.20 6.31 6.41
0.05 2.96 3.59 3.98 4.25 4.47 4.65 4.79 4.92 5.04 5.14 5.23
19
0.01 4.05 4.67 5.05 5.33 5.55 5.73 5.89 6.02 6.14 6.25 6.34
0.05 2.95 3.58 3.96 4.23 4.45 4.62 4.77 4.90 5.01 5.11 5.20
20
0.01 4.02 4.64 5.02 5.29 5.51 5.69 5.84 5.97 6.09 6.19 6.28
TABLE 5 (Continued)
The num ber of means, c
n T- c a 2 3 4 5 6 7 8 9 10 11 12
0.05 2.92 3.53 3.90 4.17 4.37 4.54 4.68 4.81 4.92 5.01 5.10
24
0.01 3.96 4.55 4.91 5.17 5.37 5.54 5.69 5.81 5.92 6.02 6.11
0.05 2.89 3.49 3.85 4.10 4.30 4.46 4.60 4.72 4.82 4.92 5.00
30
0.01 3.89 4.45 4.80 5.05 5.24 5.40 5.54 5.65 5.76 5.85 5.93
0.05 2.86 3.44 3.79 4.04 4.23 4.39 4.52 4.63 4.73 4.82 4.90
40
0.01 3.82 4.37 4.70 4.93 5.11 5.26 5.39 5.50 5.60 5.69 5.76
0.05 2.83 3.40 3.74 3.98 4.16 4.31 4.44 4.55 4.65 4.73 4.81
60
0.01 3.76 4.28 4.59 4.82 4.99 5.13 5.25 5.36 5.45 5.53 5.60
0.05 2.80 3.36 3.68 3.92 4.10 4.24 4.36 4.47 4.56 4.64 4.71
120
0.01 3.70 4.20 4.50 4.71 4.87 5.01 5.12 5.21 5.30 5.37 5.44
0.05 2.77 3.31 3.63 3.86 4.03 4.17 4.29 4.39 4.47 4.55 4.62
oo
0.01 3.64 4.12 4.40 4.60 4.76 4.88 4.99 5.08 5.16 5.23 5.29
S ource : E. S. Pearson and H. O. Hartley, Biometrika Tables for Statisticians, vol. 1 (Cambridge: Cambridge University Press, 1966).
S ource : Adapted from "Extended Tables o f the Wilcoxon Matched Pairs Signed Rank Statistics," Journal o f the American
Statistical Association 60 (1965), 864-71.
TABLE 7 L o w e r (WL) a nd U p p e r (WU) Critical Values fo r th e W ilco xo n Rank-Sum Test
α = 0 .0 2 5 fo r o n e -ta ile d te s t a n d α = 0 .0 5 fo r a tw o -ta ile d te s t
n 1: 3 4 5 6 7 8 9 10
n1: 3 4 5 6 7 8 9 10
S ource: F. Wilcoxon and R. A. Wilcox, Some Rapid Approximate Statistical Procedures (New York: American Cyanamid Company, 1964.)
n = 5 0.900 — — —
6 0.829 0.886 0.943 —
—
7 0.714 0.786 0.893
Source : E. G. Olds, 'D istribution o f Sums o f Squares o f Rank Differences for Small Samples," Annals o f Mathematical
Statistics 9(1 938).
APPENDIX B
C h a p te r 1 b.
1.2 35 is likely the estimated average age. It w ould be rather
impossible to reach all video game players.
1.4 a. The population is all m arketing managers.
b. No, the average salary was likely com puted from a sample
in order to save tim e and money.
1.6 Answers w ill vary depending on when data are retrieved. The
numbers represent tim e series data.
1.8 Answers w ill vary depending on when data are retrieved. The
numbers represent cross-sectional data.
1.10 Answers w ill vary depending on when data are retrieved. The
numbers represent cross-sectional data.
1.12 a. Qualitative
b. Quantitative, continuous
c. Quantitative, discrete
1.14 a. Ratio
b. Ordinal
c. Nominal
1.16 a. Nominal
b.
M ajor Num ber o f Students
Accounting 5
Economics 7
Finance 5
2.6 a.
Marketing 4 Response Frequency
Management 5 Good jobs 1970
Undecided 4 Affordable homes 799
c. Economics (Marketing) has the highest (lowest) num ber o f Top schools 586
students.
Low crime 1225
C h a p te r 2 Things to do 745
2.2 a.
Rating Frequency Relative Frequency
Excellent 5 0.208 b.
Good 12 0.500
Fair 4 0.167
Poor 3 0.125
Other 0.046
1,300 up to 1400 16 1
c.
2.12 This graph does not correctly depict th e data. The vertical axis
has been stretched so th a t the increase in sales appears more
pronounced than warranted.
2.14
Relative Cumulative
Class Frequency Frequency Relative Frequency 2.18 a.
Class Frequency
- 1 0 up to 0 9 0.129 0.129
—20 up to - 1 0 2
0 up to 10 31 0.443 0.572
—10 up to 0 14
10 up to 20 19 0.271 0.843
0 up to 10 13
20 up to 30 8 0.114 0.957
10 up to 20 11
30 up to 40 3 0.043 1 20 up to 30 10
a. 19 observations
14 o b s e rva tio n s are at least — 10 b u t less th a n 0.
b. 27.1%; 84.3%
b. Class Cumulative Frequency b. Cumulative
- 2 0 up to - 1 0 2 Relative Cumulative Relative
Temperature (°F) Frequency Frequency Frequency
- 1 0 up to 0 16
60 up to 70 0.061 2 0.061
0 up to 10 29
70 up to 80 0.212 9 0.273
10 up to 20 40
80 up to 90 0.424 23 0.697
20 up to 30 50
90 up to 100 0.303 33 1
40 observations are less than 20.
c. c. 9 cities
d. 42.4%; 69.7%
e.
2.20 a.
Assets
(in billions USD) Frequency
40 up to 70 9
The distribution is slightly negatively skewed.
70 up to 100 8
2.24 a.
100 up to 130 2 Vacancy Relative Cumulative Cumulative
3 up to 6 0.20 15 0.3
b. Cumulative
6 up to 9 0.40 35 0.7
Assets Relative Cumulative Relative
(in billions USD) Frequency Frequency Frequency 9 up to 12 0.20 45 0.9
c. 2 funds; 19 funds
d. 40%; 95%
e.
2.26 a.
Relative Cumulative
Age Frequency Frequency Relative Frequency
18 up to 22 45 0.45 0.45
22 up to 26 25 0.25 0.70
2.22 a.
Tem perature (°F) Frequency 26 up to 30 15 0.15 0.85
60 up to 70 2 30 up to 34 11 0.11 0.96
70 up to 80 7 34 up to 38 4 0.04 1
2.28
Stem Leaf 2.38
-8 75532000
-7 975332 1
-6 554
-5 20
7 346788
8 0123444478 The returns o f the assets are negatively correlated. The investor
w ould be wise to include these in her portfolio.
9 0001122233444445666889
10 67 2.40 a.
Utah Relative Kentucky Relative
Response Frequency Frequency
T e m pe ratu res ra nged fro m 73 to 1 0 7 . T em peratures in th e 90s
were most frequent. Yes 0.10 0.45
2.32 Spain No 0.90 0.55
3 002 b.
Netherlands
Stem Leaf
2 233455566677779
3 03559
2.42
Topamax 0.319
Lamictal 0.295
Depakote 0.135
Lyrica 0.127
Keppra 0.124
b. (0.09)(550) ≈ 50 parents. c.
2.44 a.
Region Relative Frequency
Northeast 0.165
Midwest 0.194
South 0.416
West 0.225
c.
2.46 a.
Response Frequency
2.52 a. 16%
A few days 642 b. 76%
A few long weekends 550 c.
Stem Leaf
One week 1101
3 66
Two weeks 764
4 47
b. 5 3346
6 01556779
7 013337899
C h a p te r 3
3.2 Mean = - 2.67; Median = - 3.5; Mode = - 4
3.4 Mean = 18.33; Median = 20; Mode = 15,20
3.6 a. Mean = 763.97
b. Median = 548.55. The mean does n o t reflect the typical
compensation because it is affected by outliers.
3.8 The mode.
3.22 Gg = 0.0647
3.24 a.
Time Period Growth Rate
Year 1 to Year 2 0.0667
Year 2 t o Year 3 0.0781
Year 3 to Year 4 0.1014
b. Gg = 0.082
3.26 a. Arithm etic mean = 8.605%. The return for a given year is on
average 8.61%.
3.32 a. Range = 1 0 - (- 8 ) = 18
3 .6 6
3 .6 8
d. σ = 6.07
3.34 a. Range = 12 - ( - 1 0 ) = 22
3 .7 0
d. CV = 13.16
3.36 a. 3 .7 2
c. Total return
3.40 a. In ve stm e n t B (higher mean); Investment A (smaller standard 3.76 Mean = 809.14; Median = 366; No mode. The median best
reflects typical sales since Olive Garden is clearly an outlier that
deviation)
pulls the mean up.
b. SharpeA = 1.72; SharpeB = 1.36
Investment A provides a higher reward per unit o f risk as 3.78 a. GGap= - 0.050; Gae = -0 .0 1 2
SharpeA is larger. b. American Eagle
3.42 a. Vanguard Energy Fund
3.80
b. Vanguard Energy Fund
c. SharpeEnergy = 0.45; SharpH
ealth = 0.24
The Energy Fund has a higher Sharpe ratio, meaning it offers
3.82
a higher return per u n it o f risk.
3.44 a. at least 75%
b. at least 89%
3.46 a. at least 75%
3.84
Statistic Income U nem ploym ent Debt
b. at least 89%
3.48 a. 16% Mean 74.05 9.77 983.46
b. 2.5%
c. 16%
C h a p te r 4
3.56 a. at least 75%
4.2 a. 1 to 1
b. 95%
b. 9 to 1
3 .5 8
c. 0.67 to 1
4.4 a. A Ս B = {1 ,2 ,3 ,5 ,6 } ≠ S. Thus A and B are not exhaustive. b. P(A) = 0.60
A Ո B = A ≠ Ø . T hu s A and B are n o t m utually exclusive. c. P(A Ո B) = 0.26
b. A Ս C = {1 ,2 ,3 ,4 ,6 } ≠ S . Thus A and C are not exhaustive. d. P(A | B) = P(A Ո B) / P(B) = 0.65
A Ո C ≠ Ø Thus A and C are m utually exclusive. e. P(B\AC) = P(AC Ո B)/ P (AC) = 0.35
c. A Ս D = { 1 ,2 ,3 ,4 ,5 ,6 } = S. T hus A and D are exhaustive. f. No, since P(A Ո B) = 0.26 ≠ 0.
A Ո D = Ø . T hus A and D are m utually exclusive.
g. No, since P(A | B) ≠ P(A).
d. B Ս C = { 1 ,2 ,3 ,4 ,5 ,6 } = S. Thus B and C are exhaustive.
4.34 a. IT G Total
B Ո C = {6} = Ø . Thus B and C are not m utually exclusive.
4.6 a. A and B are not exhaustive since one may not get an offer Yes 0.2214 0.3657 0.5871
from either firm , No 0.2071 0.2057 0.4129
b. A and B are not m utually exclusive since one may get an
Total 0.4286 0.5714 1
offer from both firms.
4.8 a. In 1971, P(Age ≤ 40) = 0.71. In 2006, P(Age ≤ 40) = 0.13. b. P(IT) = 0.4286
b. In 1971, P(Age ≥ 51) = 0.08. In 2006, P(Age ≥ 51) = 0.59. c. P(Yes) = 0.5871
c. Municipal managers tend to be older in 2006 compared to d. P(IT | Yes) = P(Yes Ո IT )/P(Yes) = 0.3771
1971. e. P(Yes | G) = P(Yes Ո G )/P(G) = 0.64
4.10 a. Net gain = $533.33. Net loss w ould have been $1,000. f. No, since P(IT | Yes) ≠ P(IT).
b. P(Spain wins) = 0.652. 4.36 a.
Vaccinated ( V ) Dum m y Shot (D ) Total
4.12 a. P(A Ո B) = P(A \ B)P(B) = 0.135
Infected (I ) 0.016 0.014 0.030
b. P(A Ս B) = P(A) + P(B) - P(A Ո B) = 0.815
Not Infected 0.477 0.493 0.970
c. P(B | A) = P(A Ո B)/ P (A) = 0.208
4.14 a.P(A Ո B) = 0, since A and B arem utually exclusive. Total 0.493 0.507 1
4.20 a. P(AC|BC) = P(ACՈ BC) / P(BC) = 0.48 c. B W = European woman; P(LC| BW) = 190/(310 + 190) = 0.38
b. P(ACՍ BC) = P(AC) + P(BC) - P(ACՈ BC) = 0.86 d. AM = American man; P(LC\ A M ) = 290/(210 + 290) = 0.58
c. P(A Ս B) = 1 - P (AC Ո BC) = 0.76 e. i. America: P(L \ M ) = 2 1 0/(210 + 290) = 0.42 ≠ P(L) =
(210 + 3 7 0 )/1 200 = 0.48; n o t independent
4.22 R = Reduction in unem ploym ent; E = Recession in Europe
ii. Europe: P(L \ M ) = 150/(150 + 150) = 0.50 ≠ P(L) =
a. P(RC) = 1 - P(R) = 0.82
(150 + 310)/800) = 0.58; n o t independent
b. P(RC Ո E)= P(RC| E)P(E) = 0.0752
iii. Asia:P(L \ M ) = 120/(120 + 80) = 0.60 = P(L) =
4.24 F = Foreign student; S = Smoker
(120 + 180)/500; independent
P(S | F) = P(F Ո S)/ P (F). Thus P(F) = P(F Ո S)/P (S | F) = 0.10
f. P(L | M ) = 480/(480 + 520) = 0.48 ≠ P(L) =
4.26 A = Experienced problems shopping online; B = abandoned (480 + 860)/2,500 = 0.54; n o t independent. Women
transaction or switched to a competitor's website; C = contacted tend to like the perfum e more than men.
customer-service representatives
4.40 a. P(BC) = 0.40
a. P(AC) = 0.13
b. P(A Ո B) = P(A B)P(B)
| = 0.48
b. P(A Ո B) = P(B \ A )P(A) = 0.37
P(A Ո BC) = P(A | BC)P(BC) = 0.04
c. P(A Ո C) = P(C |A)P(A ) = 0.46
c. P(A) = P(A AՈB) + P(A Ո BC)= 0.52
4.28 S = W omen face sexual harassment; T = Women use public
d. P(B\ A ) = P(A Ո B)/ P (A) = 0.9231
transportation
4.42
a. P(S Ո T) = P(S |T)P(T) = 0.23 Prior Conditional Joint Posterior
b. P(T|S) = P(S Ո T) /P(S) = 0.34 Probabilities Probabilities Probabilities Probabilities
4.30 F = Foreclosed; H = centered in Arizona, California, Florida, or P(B1) = 0.1 P(A | B1) = 0.4 P ( A Ո B 1 ) = 0.04 P (B1 | A) = 0 .0 6
Nevada
P(B2) = 0.6 P ( A | B 2 ) = 0.6 P ( A Ո B 2 ) = 0.36 P[ B 2\ A ) = 0 .5 6
P(F Ո H) = P(H \ F )P(F) = 0.0049
P(B3) = 0.3 P ( A | B 3 ) = 0.8 P ( A Ո B 3 ) = 0.24 P(B3| A) = 0.38
4.32 a.
B BC Total Total = 1 P(A) = 0.64 Total = 1
A 0.26 0.34 0.60
c. 0.1348
b. P(X > 2) = 1 - [P(X = 0) + P(X = 1)] = 0.7549
d. 0.4630
5.32 5.62
b. P(X ≤ 2) = P(X = 0) + P(X = 1) + P(X = 2) = 0.7969
c. P(X ≥ 7) = P(X = 7) + P(X = 8) = 0.0001 5.64 a. P(X = 2) = 0.0495
c. P(6 ≤ X ≤ 8) = P(- 0.17 ≤ Z ≤ 1.5) = 0.9332 - 0.4325 b. P(X > 50) = P(Z > 1.86) = 0.0314
= 0.5007 c. Women
6.30 a. P(X > 19) = P(Z > - 1 .5 ) = 1 - 0.0668 = 0.9332 6.58 a. E(X) = 15, λ = 0.0667
b. P(X > 19) = P(Z > 1.5) = 1 - 0.9332 = 0.0668 b. P(X < 10) = 1 - e-0.0667(10) = 0.4868
c. P(23 ≤ X ≤ 25) = P(0.5 ≤ Z ≤ 1.5) = 0.9332 - 0.6915 c. P(X > 25) = 1 - (1 - e -0.0667(25)) = 0.1887
= 0.2417 6.60 a. P(X < 4) = P(Z < - 1 ) = 0.1587
d. P(23 ≤ X ≤ 25) = P(3.5 ≤ Z ≤ 4.5) ≈ 1 - 0.9998 = 0.0002
b. P(Y < 4) = P(X < 1.3863) = P(Z < -1 .0 9 ) = 0.1379
6.32 a.P(10 ≤ X ≤ 20) = P( - 0 . 7 5 ≤ Z ≤ 0.5) = 0.6915 - 0.2266
c. The probability is higher if the relief tim e is normally
= 0.4649
distributed.
b. P(X > 24) + P(X < 6) = P(Z > 1) + P(Z < -1 .2 5 )
= 0.1587 + 0.1056 = 0.2643
C h a p te r 7
6.34 P(X < 28) = P(Z ≤ z) = 0.975; z = 1.96; σ = 1.53
7.2 Nonresponse bias if some people are less likely to stop at the
6.36 a.P(50 ≤ X ≤ 80) = P( - 0 . 5 ≤ Z ≤ 1) = 0.8413 - 0.3085 booth. Selection bias since the booth is only open on the
= 0.5328 weekend.
b. P(20 ≤ X ≤ 40) = P( - 2 ≤ Z ≤ - 1 ) = 0.1587 - 0.0228 7.4 a. Nonresponse bias if the people who respond are
= 0.1359 systematically different from those who do not respond.
c. P(X < x) = 0.85; z = 1.04; x = 80.8 b. Selection bias since those w ho frequent the store in the
d. P(X < x ) = 0.10; z = - 1 . 2 8 ; x = 34.4 m orning are likely to prefer an earlier opening time.
6.38 P(X ≤ 0) = 0.10; z = -1 .2 8 ; σ = 4.375 c. Selection bias since not everyone reads a newspaper.
6.40 a. Risky fund: P(X < 0) = P(Z < -0 .5 7 ) = 0.2843; Less risky Nonresponse bias since people w ho respond may be
fu n d :P(X < 0) = P(Z < - 0 .8 ) = 0.2119; Pick the less risky systematically different.
fund. 7.6 a. Both sample means w ill be norm ally distributed since the
b. Risky fund: P(X > 8) = P(Z > 0) = 0.5; Less risky fund: population is norm ally distributed.
P(X > 8) = P(Z > 0.8) = 1 - 0.7881 = 0.2119; Pick the b. Yes
riskier fund.
6.42 a. λ = 0.04; SD(X) = 25
b. (1 - e -0.04(30)) - (1 - e-0.04(20)) = 0.1481
7.8
c. (1 - e-0.04(35)) - (1 - e-0.04(15)) = 0.3022
b. P(8 < Y < 9) = P(2.08 < X < 2.20) = P(0.35 < Z < 0.50)
= 0.0547
7.12
c. P(Y < y) = 0.9; z = 1.28; x = 2.824; y = 16.84
6.48 a. λ = 0.1667; SD(X) = 6
c. Janice; her findings are more likely if a representative
b. No, the prob. o f each customer arriving is independent of sample is used.
other customers
7.14. a. The sample mean has a normal distribution because the c. The last tw o points are outside the control limits. The upward
population is norm ally distributed. trend suggests the process is becoming o u t o f control and
should be adjusted.
7.36 a.
7.16
7.18
b.
7.22
7.24 c. The control lim its have a larger spread w ith smaller sample
sizes.
c. The h ig h e r sam ple size makes less e x tre m e sam ple statistics 7.38 a.
more likely.
7.26
7.28
b.
7.44 a.
Week 5 is slightly above the control limit. The firm may w ant
to inspect the machine.
7.50
d. The one w ith the smaller sample size.
8.14
7.52
8.16
b. Yes, because the value 7 does not fall in the interval.
7.54 8.18
8.22
8.28
8.32
8 .7 4
8.36
C h a p te r 9
9.2 a. Invalid. The test is about the population parameter /j l .
b. Valid
8.38 c. Valid
d. Invalid. The null hypothesis must include some form o f the
equality sign.
c. The m a rg in o f error rises as th e c o n fid e n c e level increases. 9.4 a. Incorrect. We never accept the null hypothesis.
8.40 The population parameter o f interest is the proportion o f b. Correct.
Americans w ho support Arizona's im m igration law;
c. Incorrect. We cannot establish a claim because the null
hypothesis is not rejected.
d. Correct.
8.42 9.6 a. Type I error is to incorrectly conclude the mean w e ig h t
is different from 18 ounces. Type II error is to incorrectly
conclude the mean w e ig h t does not differ from
18 ounces.
c. The m a rg in o f e rro r in b. is h ig h e r since th e con fid e n ce level
is higher. b. Type I error is to incorrectly conclude th a t the stock price
increases on more than 60 percent o f trading days. Type II
8.44 error is to incorrectly conclude th a t the price does not
increase on more than 60 percent o f trading days.
c. Type I error is to incorrectly conclude th a t Americans sleep
less th a n 7 hours a day. Type II error is to in co rre c tly con clu de
8.46 th a t Americans do not sleep less than 7 hours a day.
9.8 a. Type I error is to incorrectly conclude th a t th e m ajority o f
8.48
voters support the candidate. Type II error is to incorrectly
conclude th a t the m ajority o f th e voters do not support the
candidate.
8 .5 0
b. Type I error is to incorrectly conclude th a t the average pizza
is less than 10 inches. Type II error is to incorrectly conclude
8 .5 2 ro u n d e d up to 139.
th a t the average pizza is not less than 10 inches.
I f D = 0 .1 2 , t h e re q u ir e d s a m p le size is 6 2.
c. Type I error is to incorrectly conclude th a t the average
8.54 ro u n d e d up to 267 ta b le t does n o t c o n ta in 250 m g. Type II e rro r is to in c o rre c tly
conclude th a t an average ta b le t contains 250 mg.
8.56
9.10 a. 3%
b. 2%
c. Fund A has a h ig h e r stan dard d e v ia tio n , th u s re q u irin g a c. Type I error is to incorrectly conclude th a t an individual has
larger sample, for the same m argin o f error. the disease. Type II error is to incorrectly conclude th a t an
individual does not have th e disease.
8 .5 8
d. We do n o t prove th a t th e individual is free o f disease if we
do no t reject the null hypothesis.
8 .6 0
9 .1 2
8 .6 2
8 .6 4
8.66 r e je c t H 0; 0 .0 2 7 4 < 0 .0 5 = α ; r e je c t H 0
9.48
c. No, since we do not reject H0.
d. z = 1.58 < 2.33 = Z0.01; we do n o t reject H0.
9.22 a. H0: µ ≤ 500,000; HA: µ > 500,000
b. p -value = 0.1681
c. No, since we do not reject H0.
9.24 a. H0: µ = 30; HA: µ ≠ 30
b. p -value = 0.0164
c. p -value = 0.0164 < 0.05 = α ; reject H0. Yes.
9.26 a. Reject H0 at α = 0.10; do not reject H0 at α = 0.01
b. Reject H0 at α = 0.10; do not reject H0 at α = 0.01
c. Reject H0 at α = 0.10; do not reject H0 at α = 0.01
d. Reject H0 at α = 0.10; do not reject H0 at α = 0.01
9.28 a. Reject H0 at α = 0.10; do not reject H0 at α = 0.01 9.50 a. z0.05 = 1.645
b. Reject H0 at α = 0 .10; do not reject H0 at α = 0.01 b. - z0.025 = - 1 .9 6 and z0.025 = 1.96
c. Reject H0 at α = 0.10; do not reject H0 at α = 0.01 c. -z0.025 1.645
=
d. Reject H0 at α = 0.10; do n o t reject H0 at α = 0.01
9 .52
9.30
c. t6 = 0.64 < 1.440 = t0.10,6; do n o t reject H0. Do n o t hire c. p -value = 0.015 < 0.05 = α ; reject H0.The concern is
another employee. supported.
d. p -value > α = 0.10; do not reject H0. 9.5 8
9.38 a. H0: µ = 1 2 ; HA : µ ≠ 12
b. No, n = 48 > 3 0 . p-value = 0.1093 > 0.05 = α . Do not reject H0.
c. t0.025,47 = 2.012; reject H0 if t47 < - 2.012 or t47 > 2.012 b. p -value = 0.1093 > 0.10 = α . Do not reject H0.
Process seems to c. Cannot conclude th a t more than 30% o f moviegoers return
to see movie for a second time.
w o rk fine.
9 .6 0
9.40
0.005 < p -value < 0.001; reject H0.The carmaker's p - v a lu e = 0 .1 0 3 8 > 0 .0 5 = α ; d o n o t re je c t H0. T h e c la im is
goal has been achieved. n o t ju s t if ie d .
9 .6 2
9 .42
0.001 < p -value < 0.005; reject H0. The mean drop o f home
p -value = 0.166 > 0 .0 1 = α ; do not reject H0. The claim is
prices is greater than 7% in San Diego.
n ot supported.
9.44 a. H0: µ = 95; HA: µ ≠ 95
9.64 a. H0: µ ≤ 10; Ha: µ -10
10.16 a. H 0 : µ 1 - µ 2 ≤ 0; H A : µ 1 - µ2 > 30
c. p - v a lu e = 0 .7 2 2 2 b. t0.05,58 = 1 .6 7 2. R e je c t H 0 i f t 58 > 1.672.
d . p - v a lu e = 0 .7 2 2 2 > 0 .0 5 = α ; d o n o t re je c t H 0. C a n n o t c. t 58 = 7 .5 8 > 1 .6 7 2; r e je c t H0. L o w -c a rb d ie te r s lo s t m o re
c o n c lu d e t h a t a v e ra g e sales d if fe r fr o m $ 1 3 0 ( th o u s a n d ). w e ig h t .
9 .7 0 1 0 .1 8
b. S in c e 0 is n o t in c lu d e d in th e in te r v a l, re je c t H 0.T h e m e a n
p -value = 0.0918 > 0.05 = α ; do n o t reject H0.
d iffe r e n c e is n o t e q u a l t o ze ro .
T he evide nce does n o t s u p p o rt th e c la im . If n = 2000,
10.20
b. -t0.05,11 = 1.796
re je c t H0. The evid e n ce s u p p o rts th e claim . c. t11 = - 1 . 7 0 2 > - 1 . 7 9 6 = - t0.05,11; d o n o t re je c t H 0. C a n n o t
9.72 a. H0: p = 0.23; HA: p ≠ 0.23 conclude th a t the mean difference is less than 0.
10.22 a. H 0 : µD ≤ 0 ; H A : µD > 0
b. H0 : µ 1 - =0; H A : µ 1 - µ 2 ≠
µ2 0;
c. The in te rv a l does n o t c o n ta in 0; re je c t H0. c. Reject H0 if t6 > 1.943 or f6 < — 1.943.
10.4 d. U = - 2 .1 0 < — 1 .9 4 3 , r e je c t Ho. A s s e rtio n is s u p p o r te d .
10.26 a. H 0 : µD = 0; H A: µD ≠ 0
d o n o t re je c t H0.
1 0 .3 0
= 0 .0 3 ± 0 .0 8 1 8 , o r [ - 0 . 1 1 1 8 , 0 .0 5 1 8 ], S in c e 0 is in th e
interval, cannot conclude th a t proportions differ.
1 0 .3 2
1 0 .3 4
p -value ≈ 0
c. R eject H0 since p -v a lu e < α = 0.05; yes
10.12 a. H 0 : µ1 - µ2 ≤ 0; H A : µ1 - µ2 > 0 b. p -value = 0
c. p -va lu e ≈ 0 < 0.05 = α ; re je ct H0. P o pu la tion
proportions differ.
1 0 .3 6
p -value = 0.0029.
c. p -value = 0.0029 < 0.05 = α ; reject H0. The proportion
b. H0: p1 - p 2 = 0; HA: p1- p2 ≠ 0
o f females th a t suffer from asthma is greater than the
c. Zero is in c lu d e d in th e in te rva l; d o n o t re je ct H0. Results do
proportion o f males.
not support study's claim.
10.48 a. H 0: p1 - p2 ≤ 0.05; HA : p1 - p2 > 0.05
10.38 a. H0: p 1 - p 2 < 0; HA'. p 1 - p 2 > 0
b. H 0: p 1 - p 2 ≤ 0 ; H A : p 1 - p 2 > 0
c. T h e w id t h o f th e in te r v a l incre a s e s w it h t h e c o n fid e n c e level.
11.6
11.10
- 1 . 9 6 < z = 0.31 < 1.96; d o n o t r e je c t H 0. C a n n o t c o n c lu d e
b. 0.05 is in clu d e d in in te rva l; c a n n o t c on clu de th a t
t h a t t h e p r o p o r t io n o f o b e s e A fr ic a n - A m e ric a n a d u lts d iffe rs
f r o m t h e p r o p o r t io n o f t h e ir C a u ca sia n c o u n te rp a r ts .
specification is being violated.
c. ±2.571
d. - 2 . 5 7 1 < t5 = - 1.01 < 2 .5 7 1 ; d o n o t re je c t H 0. C a n n o t
c o n c lu d e t h a t c ro p y ie ld w it h n e w fe r tiliz e r d iffe r s f r o m c ro p
y ie ld w it h o ld fe rtiliz e r . 11.16a. H0 : σ2 ≤ 10,000; HA : σ2 > 10,000
10.46 a. H 0: p1 - p2 ≤ 0; H A : p1 - p2 > 0
c. p -value = 0.38 c. p -value = 0.173
d. p -va lu e = 0.38 > 0.05 = α ; d o n o t re je ct H0. The realtor's d. p -value = 0.173 > 0.05 = α ; d o n o t re je ct H0. C ann ot
claim is not supported. c on clu de th a t variance in revenues is g re a te r in Asia as
Chapter 12
12.2 a. HA: At least one o f the p i (i = 1, 2, 3) differs from its
hypothesized value.
c. C a n n o t c o n c lu d e t h a t t h e s ta n d a rd d e v ia tio n d iffe rs fr o m
$ 2 0 0 fo r A n n A rb o r ; c o n c lu d e t h a t i t d iffe rs fr o m $ 2 0 0 fo r c. p -va lu e > α = 0.05; d o n o t re je c t H0. C ann ot c o n c lu d e th a t
Davis. some proportions differ from hypothesized values.
1 1 .2 0 a. 0 .0 5 1 2 .4
1 1 .2 2
1 1 .3 2 A s s u m e s a m p l i n g f r o m n o r m a l p o p u l a t i o n s .
0 .0 2 5 < p - v a lu e < 0 . 0 5 ; p - v a lu e > 0 .0 1 = α ; d o n o t r e je c t 12.20 H0: Breakup reasons and gender are independent
H 0. C a n n o t c o n c l u d e t h a t t h e v a r ia n c e e x c e e d s 5. Ha. Breakup reasons and gender are dependent
b . T h e g e n e r i c d r u g p r ic e s a r e n o r m a l l y d i s t r i b u t e d . B r e a k u p r e a s o n is
H 0. C a n n o t c o n c l u d e t h a t t h e s t a n d a r d d e v i a t i o n s o f g r o w t h
r a te s d if f e r . 12.24 a. H0: The fin a l grades are n o rm a lly d is trib u te d w ith µ = 72
and σ = 10
1 1 .4 0
HA: The final grades are not norm ally distributed w ith
µ = 72 and σ = 10
12.38 H0. H ousehold's d e lin q u e n c y in p a y m e n t is in d e p e n d e n t o f
ty p e o f heating
Class oi p i i f n o rm a l ei
HA: H ousehold's d e lin q u e n c y in p a y m e n t is d e p e n d e n t o f ty p e
Less than 50 5 0.0139 4.17 o f heating
50 up to 70 135 0.4068 122.04 re je ct H0. H ousehold's
d e lin q u e n c y in p a y m e n t is d e p e n d e n t on ty p e o f heating.
70 up to 80 105 0.3674 110.22
12.40 a. S
:H0= 0 and K = 0: HA: S ≠
0 or K ≠ 0
80 up to 90 45 0.1760 52.80
90 or above 10 0.0359 10.77 α ; d o n o t re je ct H0. C ann ot c on clu de th a t house prices are
not norm ally distributed,
b. H0: S = 0 and K = 0; HA. 5 ≠ 0 or K ≠ 0
5 up to 10 65 0.0657 15.64
d. H0: µA = µB = µC = µD ;HA: Not all population means are
10 up to 15 32 0.0697 16.59 equal
15 up to 20 38 0.0756 17.99
12.30 a. H0: p1 = 0.40, p 2 = 0.30, p 3 = 0.20, p4 = 0.10 F(3,56) = 3.58 > 2.77 = F0.05,(3,36); re je ct H0. Some p o p u la tio n
HA: At least one o f the pi (i = 1 ,2 ,3 ,4 ) differs from its means differ.
hypothesized value 13.6 a.
Source o f F c rit
V ariation SS df MS F p -v a lu e a t 10%
do n o t re je ct H0. C ann ot con clu de th a t
Between
a t least one o f th e p ro p o rtio n s differs fro m its hyp othe size d
value. Groups 548.37 5 109.67 1.37 0.250 1.96
12.32 a. H0: pA = 0.60, pB = 0.30, pC = 0.10 Within
HA: At least one o f the pi (i = A, B, C) differs from its Groups 4,321.11 54 80.02
hypothesized value
Total 4,869.48 59
T.V. s t a t io n 's
claim is n o t s u p p o rte d . b. H 0 : µ 1 = µ 2 = µ 3 = µ 4 = µ 5 = µ 6 ;H A : Not all populations
12.34 a. H0: S urviving fo r discharge is in d e p e n d e n t o f tim e o f cardiac means are equal
arrest c. p -va lu e = 0.250 > 0.10 = α , d o n o t re je ct H0. No s ig n ific a n t
HA: S urviving fo r d ischarge is d e p e n d e n t o f tim e o f cardiac differences in population means.
arrest 13.8 a.
Source o f F c rit
V a ria tio n SS df MS F p -v a lu e a t 5%
S u r v i v in g f o r d is c h a r g e is
Between
dependent o f the tim e o f the cardiac arrest.
Groups 174.72 2 87.36 2.81 0.083 3.47
12.36 H0: Effect on AD HD is in d e p e n d e n t o f tre a tm e n t
W ithin
HA: Effect on ADHD is dependent on treatm ent
Groups 652.40 21 31.07
c. p - v a lu e = 0 . 0 8 3 > 0 .0 5 = α ; d o n o t r e je c t H 0. N o s ig n if ic a n t Source o f
differences in the average w h ite n in g effectiveness o f the
V ariation SS df MS F p -va lue
detergents.
Between Groups 7,531,769.00 3 2,510,589.67 69.01 0.000
13.10 a. H0: µsnork. = µNBoard/Windsurf = µBowl = µOn-road tri =
µOff-road tri W ithin Groups 3,492,385.00 96 36,379.01
HA: Not all population mean incomes are equal. Total 11,024,154.00 99
b.
b. p -va lu e = 0 < 0.01 = α ; re je ct H0.T h e average annual
Source o f F c rit energy bills are not all th e same bv region.
V ariation SS df MS F p -value a t 5%
Between - µ2 : [ - 999.66, -6 5 6 .3 4 ]*
1
µ
Groups 4,895.15 5 979.03 37.85 0.000 2.62 µ 1 - µ3 : [- 4 4 8 .6 6 ,-1 0 5 .3 4 ]*
W ithin µ1 - µ4 : [-4 3 8 .6 6 , -9 5 .3 4 ]*
Groups 620.80 24 25.87 µ 1 - µ 3 :[379.34,722.66]*
1 3 .2 2
Total 894.67 11
in d ic a te s a diffe re n c e b e tw e e n th e tw o co rre s p o n d in g d. F(2,6) = 0.14 < 10.92 = F 0.01,(2,6) do not reject H0. Column
p o p u la tio n means means do not significantly differ.
c. µ 1 , d iffe rs fro m µ 3 ; µ 2 d iffe rs fro m µ 3 ; µ1 and µ2 d o n o t e . F(3,6) = 7 .6 3 < 9 .7 8 = F0.01,(3,6); d o n o t r e je c t H 0. R o w m e a n s
significantly differ do n o t significantly differ.
1 3 .1 6 1 3 .2 4 a.
Source o f F c rit
V ariation SS df MS F a t 1%
Rows 532.3 2 266.15 4.26 10.92
Total 1,630.7 11
Chapter 14
14.2 a.
14.4
c. do not
c. Relationship is nonlinear; see the graph in Exercise 14.10. Based on the lowest se and the highest Adjusted R2,
14.26 a. 568 the model using all 3 explanatory variables is the best
b. As x2 increases by one unit, ŷ decreases by 47.2 units, 14.46 a.
holding x1constant.
14.28 a. ŷ = 13.83 + 2.53x 1 + 0.29x2; As x2 increases by one unit, ŷ
increases by 0.29 unit, holding x1constant.
b. ŷ = 52.89
14.30 a. The positive sign on Poverty is expected, but the positive
sign on Income is unexpected.
b. As the poverty rate increases by 1%, the crime rate rises
by 53.16 crimes per 100,000 residents, holding Income
constant. The scatterplot suggests a positive correlation.
Chapter 15
14.40 15.2
b. t0.05,23 = 1.714
c. t23 = 1.67 < 1.714, we do not reject H0; β1 not positive
15.4 a. H0: β0 = 0; HA: β 0 ≠ 0; p-value ≈ 0 < 0.05 = α ; reject H0;
the intercept differs from zero
b. 0.1223 ± (2.101)(0.1794); [ − 0.2546, 0.4992]; not significant
c. Model 1; the standard error is closer to 0 and the R2
since the interval contains 0
closer to 1.
15.6 a. H0: β1 = 0 ; HA: β1 ≠ 0 ; p -value ≈ 0 < 0.05 = α ; reject H 0; x 1
14.42 a. the model and y are related
is highly unpromising
b. reported as [ − 1.67, 7.14]; no, since the interval contains 0
14.44 a. 15.8
Explanatory Adj ROA Adj Return Total Assets
b. H0: β1 = 0 ; HA: β1 ≠ 0
Variables (Model 1) (Model 2) (Model 3)
Beds 556.89
c. The confidence interval is narrower because it assumes that
(0.9783)
the expected value of the error term is zero, whereas the
Baths 92022.91* prediction interval incorporates the nonzero error term.
(0.0009) 15.30 a. Regress y on x* = x − 80 to get ŷ0 = 1045.5861, se(ŷ0) =
Se 74984.98 14.3802, and se = 63.2606; 1045.5861 ± (2.064)(14.3802);
[1015.91, 1075.27]
R2 0.7237
F 27.9348
(p-value) (0 .0000) 15.32 a. H0: β j = β 2 = 0; HA: At least one β1 ≠ 0; p-value ≈ 0 < 0.05
= α ; x1and x2 jointly significant
*NOTES: Param eter estim ates are at th e to p w ith th e p -values in parentheses; b. H0: βj = 0; HA: βj + 0;
* represents significance at 5% level. G oodness-of-fi t measures are a t th e b o tto m . p-value = 0.3832 > 0.05 = α ; do not reject H0; x1not
b. H0 : β1= β2 = 0; HA :At least one β j ≠ 0; p-value ≈ 0 < 0.05 significant
= α ; reject H0; explanatory variables jointly significant p-value = 0.3666 > 0.05 = α ; so do not reject H0; x 2 not
c. H0 : β1= 0; HA : βj ≠ 0; All coefficients have an associated significant
p-value < α = 0.05, except for β 2 only Beds not significant c. Multicollinearity is likely since both explanatory variables
are individually insignificant even though they are jointly
15.14
significant. (Also, is high).
Standard
15.34 a.
Coefficients Error t Stat p-value
Intercept 1.5804 0.9296 1.7002 0.0945
15.22
R estricted R estricted
Variable Unrestricted (β1 = β2 = 0) (β3 = β4 = 0)
Chapter 16
16.2 x = 10,15: ŷ = 92,98 (linear); ŷ = 161,96.5 (quadratic);
ŷ = 120, 115 (cubic)
16.4 a.
15.46
Variable Coefficients
Intercept − 33.40*
( 0 . 01 )
P/E 3.97*
( 0 . 00 )
P/S − 3.37 ŷ = 1.1006 + 0.5828x; ŷ = 12.1338 − 3.1034x + 0.2565x2
(0 .2 1 ) b. The quadratic regression is better because of the higher
Se 13.64 adjusted R2(0.78 > 0.42)
R2 0.40 c. x = 4, ŷ = 3.82; x = 6, ŷ = 2.75; x = 12, ŷ = 11.83
F 9.10
(p-value) ( 0 .00 )
16.6 a. x = 2,3: ŷ = 22.5,23.85 (linear); ŷ = 21.84,21.79
N OTES: Parameter estimates are at the top w ith the p-values in parentheses; (quadratic); ŷ = 21.91,21.57 (cubic)
're p re se n ts significance at 5% level. G oo d ne ss-o f-fit measures are a t th e b o tto m . b. The quadratic m odel is the best (highest adjusted R2)
b. H0: = β 2 = 0; HA: At least one β1 ≠ 0; p-value ≈ 0 < 0.05 16.8
= α ; reject H0; P/E and P/S jointly significant b. Size2 is significant (p-value = 0.01 < 0.05 = α ) , and
c. P/E: p-value ≈ 0 < 0.05 = α ; reject H0; P/E significant sample regression with Size2 has greater adjusted R2
P/S: p-value = 0.21 > 0.05 = α ; do not reject H0, P/S not (0.0959 > 0.0490)
significant
c. ŷ = 96.2515 − 4.4075 ln(x) ; i f x decreases from 1000 to
500, ŷ increases from 65.81 to 68.86, th a t is, by 3.05 years.
d. Since R2 = 0.6717 > 0.4726, the logarithm ic m odel is better.
e. Life expectancy increases proportionately more for females
16.10 a. M odel 1: As x increases by one unit, ŷ decreases by 4.2 units. with a decrease in the people-to-physician ratio; females
Model 2: As x increases by one percent, ŷ decreases by are likely to benefit more.
about 2.8 units. 16.22 a. Age with
Model 3: As x increases by one unit, ŷ decreases by 4%. R2 = 0.3931. For the exponential m odel, we com pute R2 in
Model 4: As x increases by one percent, ŷ decreases by 0.8 %. term s o f y for comparison w ith th e linear model. Using the
predicted reported by Excel, we transform it in to ŷ and
b. ŷ decreases by 4.2 (model 1); ŷ decreases by 2.8
then com pute ryŷ = 0.5147. Thus, R2 = 0.51472 = 0.2649.
(Model 2); ŷ decreases by 3.22, or 3.92% (Model 3);
The linear m odel is better because o f a higher R2 (0.3931 >
ŷ decreases by 0.59, or 0.78% (Model 4).
0.2649).
16.12
b. H0: β 3 = β 4 = 0; H A: At least one β j ≠ 0
c. do
n o t reject H0; Age and Age2 n o t jo in tly significant
16.24
16.14 a.
Chapter 17
17.2 a. ŷ = 160 + 15(1) + 32(1) = 207
b. ŷ = 1 6 0 + 15(0) + 32(0) = 160
17.4 a. If d changes from 0 to 1, w ith x held constant, ŷ increases by
about 15%
f ytȳ ŷt et = yt − ŷt
17.32 Relevant regression results from Minitab: 1 14 — — —
1 1 1 8.37 —
2- Nov 1193.57 1191.97 — —
1 2 2 12.78 — —
3- Nov 1197.96 1204.2 — —
1 3 3 8.84 11.69 0.7564
4-Nov 1221.06 1214.96 1191.97 29.09
⋮ ⋮ ⋮ ⋮ ⋮ ⋮
⋮ ⋮ ⋮ ⋮ ⋮
5 2 18 10.58 11.59 0.9130
30-Nov 1180.55 — 1191.84 -1 1 .2 9
5 3 19 13.35 — —
1-Dec 1185.90
5 4 20 19.77 — —
t yt et = yt −ŷt
A t ŷt Quarter Quarter 1 Quarter 2 Quarter 3 Quarter 4
1-Nov 1184.38 1184.38
Unadjusted 0.7412 0.8995 0.8145 1.4629
2-Nov 1193.57 1188.06 1184.38 9.19
A djusted 0.7567 0.9183 0.8315 1.4935
⋮ ⋮ ⋮ ⋮ ⋮
30-Nov 1180.55 1186.30 1190.14 − 9.59 First quarter series is 24.33% below its average quarterly level;
1-Dec 1186.30 fourth quarter series is 49.35% above its average quarterly
level.
18.22 b. Calculating the ratio-to-moving averages:
c. Exponential smoothing method
d. Yes, the December 1 absolute errors induced by the Year Quarter t y ȳ y/ ȳ
methods are 20.17 and 19.77, respectively. Year 1 Quarter 1 1 6.49 — —
18.10 a. ŷ 21 = 13.54 + 1.08(21) = 36.22
Year 1 Quarter 2 2 7.34 — —
b. ŷ 21 = 18.26 + 0.92(21) − 0.01 (21)2 = 33.19
Year 1 Quarter 3 3 7.11 8.01 0.8878
⋮ ⋮ ⋮ ⋮ ⋮ ⋮
18.12 a. From the graph, a quadratic trend model seems to be
appropriate. Year 5 Quarter 2 18 10.08 10.78 0.9353
b. ŷt = 11.7606 + 0.3144t; adjusted R2 = 0.8356 Year 5 Quarter 3 19 9.78 — —
ŷt = 10.0651 + 0.7768t − 0.0220t2; adjusted R2 = 0.9481 Year 5 Quarter 4 20 14.88 — —
Quadratic trend model is preferred since its adjusted R2 is
higher.
Seasonal indices: Seasonal indices:
Quarter Quarter 1 Quarter 2 Quarter 3 Quarter 4 Month Jan Feb Mar Apr May Jun
Unadjusted 0.8484 0.9345 0.8880 1.3278 Unadjusted 0.9582 0.9935 0.9852 1.0100 1.0243 1.0595
Adjusted 0.8487 0.9348 0.8883 1.3282 Adjusted 0.9550 0.9902 0.9819 1.0066 1.0209 1.0559
c. T t= 5.160547 − 0.02228f
d.
Year Month t ŝt Tt ŷ t = Tt × ŝ t
d. The estimated exponential trend m odel w ith dumm ies has
a lower MSE. Forecasts: 2011 Jan 61 0.9550 3.8017 3.6306
b. Estimated exponential trend model with seasonal dummies: c. The linear trend model outperforms the exponential model.
18.30 a. Estimated linear trend model with seasonal dummies:
ŷ t = 1.4702 + 0.0174d1 − 0.0729d2 − 0.0630d3 − 0.0137t;
adjusted R2 = 0.8560;
b. Estimated quadratic trend model with seasonal dummies:
ŷ t = 1.4155 + 0.0174d 1 − 0.0734d2 − 0.0636d3 − 0.0040t
− 0.0003t2; adjusted R2 = 0.8800
c. The quadratic model is used for 2009 forecasts as it has
The exponential model has lower MSE. Forecasts:
greater adjusted R2.
Year Month d1 d2 t ŷt
Year Quarter ŷt
6 Jan 1 0 61 7186.15
2009 1 0.9802
6 Feb 0 1 62 6679.55
2009 2 0.8656
18.40
The estimated
exponential trend model has lower MSE, and it yields the
forecast ŷ 37 = 3,189.63 billion.
19.4
c. ŷt = −37.2161 + 1.0961yt−1; MSE = 4104.42. T he e s tim a te d
AR(1) m odel has the lowest MSE, and it yields the forecast
ŷ 37 = 3,913.21 billion.
18.46 a. The graph shows a relatively fla t trend w ith a succession o f
alm ost identical spikes in July o f each year, indicative o f the
presence o f a strong seasonality in the series.
19.6 b.
Simple Price index Updated Price
b. $13,901 Year (Base = 2004) Index (Base = 2007)
19.8 2004 100.00 82.70
Date Adjusted Close Price Monthly Return
2005 108.12 89.42
Dec-10 $21.48 0.0238
2006 115.70 95.69
Nov-10 $20.98 0.0634
2007 120.91 100.00
Oct-10 $19.73 —
2008 126.96 105.00
2009 132.02 109.19
19.10 a.
Year Price Price Index
c. Tuition increased by 20.91 percent from 2004 through 2007,
1994 62 100.00 but only 9.19 percent from 2007 through 2009.
1995 60 96.77 19.18 a. Relative to 2007, the 2009 prices of omelet, pancake,
and cereal increased by 5.26%, 28.57%, and 21.43%,
1996 64 103.23
respectively.
1997 67 108.06
Simple Price Index
1998 66 106.45
Year Omelet Pancakes Cereal
1999 70 112.90
2007 100.00 100.00 100.00
2000 74 119.35
2008 110.53 121.43 114.29
2001 72 116.13 2009 105.26 128.57 121.43
2002 70 112.90
b. Relative to 2007, the prices of the three breakfast items
increased by 14.89% and 17.02% in 2008 and 2009,
b. Increase of 6.45 percent.
respectively.
19.12 a. Σpit Unweighted Aggregate
Simple Price Index
Year Price Index
Year Product 1 Product 2 Product 3
2007 11.75 100.00
2008 100.00 100.00 100.00
2008 13.50 114.89
2009 104.20 98.56 102.21
2009 13.75 117.02
2010 108.39 99.28 98.90
19.20 a.
Simple Price Index
b. Product 1's price increased each year, while product 2's price
decreased each year. Product 3's price increased in 2009, but Region 2007 2008 2009
then decreased in 2010. Northeast 100.00 94.24 83.55
19.14 a.
Midwest 100.00 93.25 88.29
Month Jan Feb Mar Apr May Jun South 100.00 94.74 86.47
Price 3.25 3.18 3.56 3.82 3.97 4.48 West 100.00 80.61 65.46
Simple
b. Home prices dropped significantly; more drastically in the
Price Index 100.00 97.85 109.54 117.54 122.15 137.85 West.
19.22
Nominal Value Price Index Real Value
Month Jul Aug Sep Oct Nov Dec
32 100 32.00
Price 4.46 4.16 3.79 3.39 2.46 1.82 37 102 36.27
Simple
Price Index 39 103 37.86
137.23 128.00 116.62 104.31 75.69 56.00
42 108 38.89
b. The price rose by 37.85 percent.
19.24
19.16 a.
Year Tuition Simple Price Index (Base = 2004)
2004 $36,850 100.00
19.26 a-b.
2005 $39,844 108.12 Year Nominal Value Price Index Real Value
2009 38 100 38.00
2006 $42,634 115.70
2010 40 103 38.83
2007 $44,556 120.91
2011 42 112 37.50
2008 $46,784 126.96
2009–2010: Nominal values increase by 5.26 percent. Real
2009 $48,650 132.02 values increase by 2.18 percent.
2010–2011: Nominal values increase by 5 percent. Real b. Year Unweighted Aggregate Price index
Σ pit
values decrease by 3.43 percent.
2009 177 100.00
2010 180 101.69
The 2009 starting salary must be reduced to $89,156(0.9965) = 2009 111.0 172.9 64.20
$88,844.
b. Net CPI Real
19.34 a.
Adjusted Simple Price Index
Year Income (1982–84 = 100) Net Income
Date Close Price (Base = Oct 09)
2006 21.2 201.59 10.52
0ct-09 78.89 100.00
Nov-09 78.54 99.56 2007 3.6 207.34 1.74
d. Increase by 3.33 percent. February, 2008 8.29 − 0.0727 212.70 0.0023 − 0.0748
19.36 a. March, 2008 6.01 − 0.2750 213.54 0.0039 − 0.2779
Simple Price Index (Base 2009)
Year Product 1 Product 2 Product 3
2009 100.00 100.00 100.00 Chapter 20
20.2 a. Reject H0 if T ≤ TL = 1
2010 105.26 97.87 106.67
b. T = T + = 3 > 1 = Tl, do not reject H0. Cannot conclude
2011 110.53 104.26 124.44 th a t population median is less than 150.
20.4 a. H0: m ≥ 10; HA : m < 10 20.26 a. H0: m 1 = m 2 = m 3 = m 4; HA. Not all population medians
b. T =T+ = 1 are equal
c. Reject H0 if T ≤ TL = 2
d. T = 1 < 2 = TL, reject H0. Population median is less than 10. c. With d f = 3, p -value is strictly between 0.01 and 0.025.
20.6 a. H0: m ≤ 25; H A : m > 25 d. p -value > α = 0.01; do not reject H0. Cannot conclude that
b. T = T+ = 35; T is assumed normally distributed with
some population medians differ.
20.28 a. H0: m 1 = m 2 = m 3; HA: Not all population medians are equal
b. W ith d f = 4 , p -value is strictly between 0.01 and 0.025. d. p -va lu e = 0.2938 > 0.05 = α ; d o n o t re je ct H 0. No.
c. p -value < 0.05 = α ; reject H0. Some population medians 20.50 a. H0: ρ = 0.50; HA: ρ ≠ 0.50
differ.
c. Reject H0 i f z < − 1.645 o r z >1.645. 20.62 a. H0: m D = 0; H A: m D ≠
0
d. − 1.645 < z = − 1.26 < 1.645; do not reject H0. No. b. T = T+ = 9.5
20.52 a. H0: The elements occur random ly; H A: The elements do not c. Reject H0 if T ≤ TL = 0 or T TU
≤ = 21
occur randomly. d. 0 < T = 9.5 < 21; do not reject H0. Population median
difference is not significantly different from 0. Farmer
should not be concerned.
20.64 a. H0: m A − m B = 0; H A: m A − m B + 0
c. p -value = 0.0128 b. Given n 1 = n 2 = 40, W = W1, = 1492
d. p -value = 0.0128 < 0.05 = α ; reject H0. Yes.
20.54 a. H0: outcomes occur random ly; H A: Outcomes do not occur
randomly
c. Reject H0 i f z < − 2.58 o r z > 2.58 c. p -value ≈ 0 < 0.05 = α ; reject H0. Yes.
20.70 H0: GDP grow th rate is random; H A: GDP grow th rate is not
d. − 2.58 < z = − 1.72 < 2.58; do not reject H0. Yes.
random.
20.58 H0; GDP grow th rate is random ; H A: GDP grow th rate is not
random.
GLOSSARY
least one of these events occurs, is equal to the probability that over- or underestimate a population parameter.
A occurs, plus the probability that B occurs, minus the prob Binary choice models Regression models that use a dummy
ability that both A and B occur, or equivalently, (binary) variable as the response variable. Also called discrete
P (A ᑌ B ) = P(A ) + P(B ) − P(AᑎB ). choice or qualitative response models.
Binomial probability distribution A description of the
Adjusted close price Stock price data adjusted using appro probabilities associated with the possible values of a binomial
priate dividend and split multipliers. random variable.
Adjusted R 2 A modification of the coefficient o f determi Binomial random variable The number of successes
nation that explicitly accounts for the sample size n and the achieved in the n trials of a Bernoulli process.
number of explanatory variables k in regression analysis; often
Block In ANOVA, a matched group of observations from
used as a criterion for selecting among regression models with
each population. In general, blocks are the levels at which the
different numbers of explanatory variables.
extraneous factor is held fixed, so that its contribution to the
Aggregate price index A representation of relative price total variation of the data can be measured.
movements for a group of items.
Box plot A graphical display of the smallest value, quartiles,
Alternative hypothesis (H A) In a hypothesis text, the alter and the largest value of a data set.
native hypothesis contradicts the default state or status quo
Capital gains yield The gain or loss resulting from the in
specified in the null hypothesis. Generally, whatever we wish
crease or decrease in the value o f an asset.
to establish is placed in the alternative hypothesis.
Causal forecasting models Quantitative forecasts based on
Analysis of variance (ANOVA) A statistical technique used
a regression framework, where the variable o f interest is related
to determine if differences exist between several population
to a single or multiple explanatory variables.
means.
Central limit theorem (CLT) The CLT states that the sum
Annualized return A measure equivalent to the geometric
or mean of a large number of independent observations from
mean return.
the same underlying distribution has an approximate normal
A priori probability A probability value based on logical distribution. The approximation steadily improves as the num
analysis rather than on observation or personal judgment. ber of observations increases.
Arithmetic mean The arithmetic mean is the primary mea Chebyshev’s theorem For any data set, the percentage of ob
sure of central location. It is calculated by adding up the values servations that lie within k standard deviations from the mean
of all the data points and dividing by the number of data points will be at least 1 − 1/ k 2, where k is any number greater than 1.
in the population or sample.
Chi-square (χ 2) distribution A family of distributions
Autoregressive model In forecasting, a model where where each distribution depends on its particular degrees of
lagged values of the response variable are used as explanatory freedom df. It is positively skewed, but becomes increasingly
variables. symmetric as d f increase.
Average growth rate For growth rates g 1, g 2, . . . , g n, the Classes Intervals of quantitative data.
average growth rate Gg is computed as
Cluster sampling A population is first divided up into mutu
ally exclusive and collectively exhaustive groups of observa
tions, called clusters. A cluster sample includes observations
where n is the number of multi-period growth rates.
from randomly selected clusters.
Balanced data A completely randomized ANOVA design
Coefficient of determination (R 2) The proportion of the
with an equal number of observations in each sample.
sample variation in the response variable that is explained by
Bar chart A graph that depicts the frequency or relative fre the sample regression equation; used as a goodness-of-fit mea
quency of each category of qualitative data as a bar rising verti sure in regression analysis.
cally from the horizontal axis.
Coefficient of variation (CV) The ratio of the standard devi
B ayes’ theorem The rule for updating probabilities is ation of a data set to its mean; a relative measure of dispersion.
Combination formula The number of ways to choose x ob
jects from a total of 11 objects, where the order in which the x
objects is listed does not matter, is
where P (B ) is the prior probability and P (B |A ) is the posterior
probability.
Bernoulli process A series of n independent and identical Complement rule The probability of the complement of an
trials of an experiment such that each trial has only two possible event is one minus the probability of the event, or equivalently,
outcomes, conventionally labeled success and failure, and each P(A c) = 1 − P(A ).
time we repeat the trial, the probabilities of success and failure
Completely randomized design An experiment in which
remain the same.
independent random samples are drawn from the c different
Between-treatments variability In ANOVA, a measure of levels of a given factor (c populations), ignoring all other pos
the variability between sample means. sible factors.
Conditional probability The probability of an event given Decomposition analysis A method of estimating trend and
that another event has already occurred. seasonal components from a time series and then recomposing
Confidence coefficient The probability that a given confi them to make forecasts.
dence interval will contain the population parameter of interest. Deflated time series A series obtained by adjusting a given
Confidence interval A range of values that, with a certain mea series for changes in prices, or inflation. The two most com
sure of confidence, contains the population parameter of interest. monly used price indices for deflating economic time series are
the Consumer Price Index, CPI, and the Producer Price Index,
Consistency An estimator is consistent if it approaches the
PPI.
unknown population parameter being estimated as the sample
size grows larger. Degrees of freedom The number of independent pieces of
information that go into the calculation of a given statistic.
Consumer price index (CPI) A monthly weighted aggregate
Many probability distributions are identified by the degrees of
price index, computed by the U.S. Bureau of Labor Statistics,
freedom.
based on the prices paid by urban consumers for a representa
tive basket o f goods and services. Dependent events The occurrence o f one event is related to
the probability of the occurrence of the other event.
Contingency table A table that shows frequencies for two
qualitative or categorical variables, x and y, where each cell Descriptive statistics The summary of a data set in the form
represents a mutually exclusive combination o f the pair o f x of tables, graphs, or the calculation of numerical measures.
and y values. Deterministic relationship In regression analysis, a rela
Continuous (random) variable A variable that assumes (in tionship in which the value of the response variable is uniquely
finitely) uncountable values and can take on any value within determined by the values of the explanatory variables.
an interval. Discrete (random) variable A variable that assumes either a
Continuous uniform distribution A distribution describing finite number of values or an infinite sequence of values.
a continuous random variable that has an equally likely chance Discrete uniform distribution A symmetric distribution
o f assuming a value within a specified range. where the random variable assumes a finite number of values
Control chart A plot of statistics of a production process and each value is equally likely.
over time. If the statistics randomly fall in an expected range, Dummy variable A variable that takes on values of 0 or 1.
then the production process is in control. If the statistics reveal It is commonly used to describe a qualitative variable with two
an undesirable trend, then adjustment of the production process categories.
is likely necessary. Dummy variable trap A linear regression model where the
Correlation coefficient A measure that describes the di number o f dummy variables equals the number of categories
rection and strength of the linear relationship between two o f a qualitative variable; in a correctly specified model, the
variables. number of dummy variables is one less than the number of
Covariance A measure that reveals the direction of the linear categories.
relationship between two variables. Efficiency An unbiased estimator is efficient if its standard
Critical value In a hypothesis test, the critical value is a deviation or standard error is lower than that of other unbiased
point that separates the rejection region from the non-rejection estimators.
region. Empirical probability A probability value based on observ
Cross-sectional data Values of a characteristic of many sub ing the relative frequency with which an event occurs.
jects at the same point in time or without regard to differences Empirical rule Given a sample mean a sample stan
in time. dard deviation s, and a relatively symmetric and bell-shaped
Cubic regression model In regression analysis, the relation distribution:
ship between the explanatory variable and the response vari • Approximately 68% of all observations fall in the interval
able is modeled as a polynomial of order 3; it is used when the
influence of that explanatory variable on the response variable
• Approximately 95% of all observations fall in the interval
has two changes o f sign.
and
Cumulative distribution function A description o f the
• Almost all observations fall in the interval
probability that the value of a random variable X is less than or
equal to a particular value x, P (X ≤ x ). Endogeneity In regression analysis, a situation where the
Cumulative frequency distribution A distribution of quan error term is correlated with the explanatory variables.
titative data recording the number of observations that falls Error sum of squares (SSE) In ANOVA, a measure o f the
below the upper limit of each class. degree o f variability that exists even if all population means are
Cumulative relative frequency distribution A distribution the same. Also known as within-sample variation. In regression
of quantitative data recording the fraction or proportion of ob analysis, it is the sum of the squared residuals.
servations that falls below the upper limit of each class. Estimate A particular value of an estimator.
Cyclical component Wave-like fluctuations or business cy Estimator A statistic used to estimate a population parameter.
cles of a time series, often caused by expansion and contraction
Event A subset of the sample space.
of the economy.
Exhaustive events When all possible outcomes of a random Grand mean In ANOVA, the sum of all observations in a
experiment are included in the events. data set divided by the total number of observations.
Expected return of a portfolio A weighted average of the Heteroskedasticity In regression analysis, a violation of the
expected returns o f the individual assets in the portfolio. assumption where the variance of the error term is not the same
Expected value A weighted average of all possible values of for all observations.
a random variable. Histogram A graphical depiction of a frequency or relative
Experiment A trial that results in any one of several possible frequency distribution; it is a series of rectangles where the
outcomes. width and height o f each rectangle represents the class width
and frequency (or relative frequency) of the respective class.
Explanatory variables In regression analysis, the variables
that we assume affect the response variable. They are also Hypergeometric probability distribution A description of
called the independent variables, predictor variables, control the probabilities associated with the possible values of a hyper
variables, or regressors. geometric random variable.
Exponential probability distribution A continuous prob Hypergeometric random variable The number of suc
ability distribution used to model lifetimes or failure times. It is cesses achieved in the n trials of a Bernoulli-type process in
based on one parameter that is referred to as the rate parameter. which the trials cannot be assumed to be independent.
Exponential regression model A semi-log regression model Hypothesis test A statistical procedure to resolve conflicts
in which only the response variable is transformed into natural between two competing opinions (hypotheses) on a particular
logs. population parameter of interest.
Exponential smoothing In time series analysis, a smoothing Income yield The direct cash payments from an underlying
technique that assigns exponentially decreasing weights as the asset, such as dividends, interest, or rental income.
observations get older. Independent events The occurrence of one event does not
Exponential trend model A model used for a time series affect the probability of the occurrence o f the other event.
that grows by an increasing amount each time period. Independent random samples Two (or more) random sam
Factorial formula The number of ways to assign every ples are considered independent if the process that generates
member of a group of size n to n slots is n ! = n × (n − 1) X one sample is completely separate from the process that gener
(n − 2) × (n − 3) × . . . × 1. ates the other sample.
F distribution A family of distributions where each distribu Index number An easy-to-interpret numerical value that
tion depends on two degrees o f freedom: the numerator degrees reflects a percentage change in price or quantity from a base
of freedom d f1 and the denominator degrees of freedom df2. It value. The most common example is a price index.
is positively skewed, but becomes increasingly symmetric as Inferential statistics The practice of extracting useful infor
d f 1and df2 increase. mation from a sample to draw conclusions about a population.
Finite population correction factor A correction factor that Inflation rate The percentage rate of change of a price index
accounts for the added precision gained by sampling a larger over time.
percentage of the population. It is recommended when the sam In-sample criteria Calculations showing how well a fore
ple constitutes at least 5% o f the population. casting model predicts values within a sample of data.
Fisher equation A theoretical relationship between nominal Interquartile range (IQR) The difference between the quar
returns, real returns, and the expected inflation rate. tile values Q3 and Q 1.
Fisher’s least difference (LSD) method In ANOVA, a test Interval estimate See Confidence interval.
that determines which means significantly differ by computing
Interval-scaled data Values of a quantitative variable that
all pairwise differences of the means.
can be categorized and ranked, and in which differences be
Frequency distribution A grouping of qualitative data into tween values are meaningful.
categories, or quantitative data into intervals called classes, re
Inverse transformation A standard normal variable Z can be
cording the number of observations that fall into each category
transformed to the normally distributed random variable X with
or class.
mean µ and standard deviation σ as X = µ + Z σ .
Geometric mean return For multiperiod returns R 1 R 2, . . . ,
Investment return The net gain or loss in value of an invest
Rn, the geometric mean return GR is computed as
ment over a time period.
Jarque-Bera test A test, using the chi-square statistic, to de
where n is the number of multiperiod returns. termine if sample data are derived from the normally distrib
Goodness-of-fit test A test, using the chi-square statistic, to uted population.
determine if the sample proportions resulting from a multino Joint probability The probability o f the occurrence of two
mial experiment differ from the hypothesized population pro events, A and B.
portions specified in the null hypothesis. Joint probability table A contingency table whose frequen
Goodness-of-fit test for normality A test, using the chi- cies have been converted to relative frequencies.
square statistic, to determine if sample data are derived from Kruskal-Wallis test A nonparametric test to determine
the normally distributed population. whether differences exist between several population medians.
Kurtosis coefficient A measure of whether data is more or cell p i remains the same; the sum of the cell probabilities is one,
less peaked than a normal distribution. p1 + P 2 + ... + Pk = 1.
Laspeyres price index A weighted aggregate price index Multiple regression model In regression analysis, more
based on quantities evaluated in the base period. than one explanatory variable is used to explain the variability
Linear probability model (LPM) A linear regression model in the response variable.
applied to a binary response variable. Multiplication rule The probability that A and B both occur,
Linear trend model A regression model used for a time or a joint probability, is equal to the probability that A occurs
series that grows by a fixed amount each time period. given that B has occurred times the probability that B occurs,
that is,
Logarithmic regression model A semi-log regression
model in which only the explanatory variable is transformed P (A ᑎ B ) = P(A |B)P(B).
into natural logs.
Logit model A regression model that ensures that the pre Mutually exclusive events Events that do not share any
dicted probability o f the binary response variable falls between common outcome o f a random experiment.
zero and one. Negatively skewed (left-skewed) distribution A distribu
Log-log regression model A regression model in which both tion in which extreme values are concentrated in the left tail
the response variable and the explanatory variable are trans of the distribution.
formed into natural logs. Nominal data Values o f a qualitative variable that differ
Lognormal probability distribution A positively skewed merely by name or label.
continuous probability distribution used to model variables that Nominal return Investment return that makes no allowance
are known to be positively skewed. for inflation.
Marginal (unconditional) probability The probability of N on-causal forecasting m odels Quantitative forecasts
an event without any restriction. that do not present any explanation o f the m echanism gen
Margin of error A value that accounts for the variability of erating the variable o f interest and simply provide a method
the estimator and the desired confidence level o f the interval. for projecting historical data. Also called purely time-series
Often the confidence interval is specified as: Point Estimate ± models.
Margin of Error. Nonparametric tests Statistical tests that rely on fewer as
M atched-pairs sample W hen samples are matched or sumptions concerning the distribution of the underlying pop
paired in some way. The first type o f matched-pairs sample ulation. Often used when the underlying distribution is not
is characterized by a measurement, an intervention o f some normal and the sample size is small.
type, and then another measurement. The second type of Nonresponse bias A systematic difference in preferences
matched-pairs sample is characterized by a pairing o f obser between respondents and nonrespondents of a survey or a poll.
vations, where it is not the same individual or item that gets Normal curve A graph depicting the normal probability den
sampled twice. sity function.
Mean Often used as a short form for the arithmetic average. Normal distribution The most extensively used probabil
It is calculated by adding up the values of all the data points ity distribution in statistical work and the cornerstone of sta
and dividing by the number of data points in the population tistical inference. Also called the symmetric or bell-shaped
or sample. d istrib u tio n .
Mean absolute deviation (MAD) In forecasting, the mean Normal transformation Any normally distributed random
of the absolute residuals; it is used for model selection. variable X with mean µ. and standard deviation σ can be trans
Mean square error (MSE) In forecasting, the residual sum formed into the standard normal random variable Z as
of squares divided by the number o f observations; it is used for
model selection.
Mean-variance analysis The idea that we measure the per Null hypothesis (H0) In a hypothesis test, the null hypothesis
formance of an asset by its rate of return and evaluate this rate corresponds to a presumed default state of nature or status quo.
of return in terms o f its reward (mean) and risk (variance).
Ogive A graph of the cumulative frequency or cumulative
Median The middle value o f a data set. relative frequency distribution in which lines connect a series
Mode The most frequently occurring value in a data set. of neighboring points, where each point represents the upper
Moving average method In time series analysis, a smooth limit of each class and its corresponding cumulative frequency
ing technique based on computing the average from a fixed or cumulative relative frequency.
number m o f the most recent observations. One-tailed hypothesis test A test in which the null hypoth
Multicollinearity In regression analysis, a situation where esis is rejected only on one side o f the hypothesized value of
two or more explanatory variables are linearly related. the population parameter.
Multinomial experiment A series of n identical trials of a One-way ANOVA A statistical technique that analyzes the
random experiment, such that for each trial: there are k pos effect o f one categorical variable or factor on the mean.
sible outcomes or categories, called cells; each time the trial is Ordinal data Values of a qualitative variable that can be cat
repeated, the probability that the outcome falls into a particular egorized and ranked.
Ordinary least squares (OLS) method A regression tech Prediction interval In regression analysis, an interval that
nique for fitting a straight line that is “closest” to the data. Also pertains to the individual value of the response variable defined
known as the method of least squares. for specific explanatory variables.
Outliers Extreme small or large values. Probability A numerical value between 0 and 1 that mea
bar chart A control chart that monitors the proportion of sures the likelihood that an uncertain outcome occurs.
defectives (or some other characteristic) of a production process. Probability distribution Every random variable is associ
Paasche price index A weighted aggregate price index based ated with a probability distribution that describes the variable
on quantities evaluated in the current period. completely. It is com mon to define discrete random vari
ables in terms o f their probability mass function and con
Parameter A constant characteristic of a population.
tinuous random variables in terms of their probability density
Partial F test See test o f linear restrictions. function.
Percentiles The p th percentile divides a data set into two Producer price index (PPI) A monthly weighted aggre
parts: Approximately p percent of the observations have values gate price index, com puted by the U.S. Bureau o f Labor
less than the p th percentile; approximately (100 − p ) percent Statistics, based on prices m easured at the wholesale or pro
of the observations have values greater than the p th percentile. ducer level.
Permutation formula The number of ways to choose x ob p -Value In a hypothesis test, the likelihood o f observing a
jects from a total of n objects, where the order in which the x sample mean that is at least as extreme as the one derived
objects is listed does matter, is from the given sample, under the assumption that the null
Pie chart A segmented circle portraying the categories and hypothesis is true.
relative sizes of some qualitative variable. Quadratic regression model In regression analysis,
Point estimate The value of the point estimator derived from the relationship between the explanatory variable and the
a given sample. response variable is modeled as a polynomial o f order 2; it
is used when the influence o f that explanatory variable on
Point estimator A function of the random sample used to make
the response variable changes sign (U-shaped or inverted
inferences about the value of an unknown population parameter.
U-shaped curve).
Poisson process A random experiment in which the number
Qualitative forecasting Forecasts based on the judgment of
of successes within a specified time or space interval equals
the forecaster, using prior experience and expertise.
any integer between zero and infinity; the numbers of successes
counted in nonoverlapping intervals are independent from one Qualitative variable A variable normally described in
another; and the probability that success occurs in any interval words rather than numerically; it is often denoted by a binary
is the same for all intervals of equal size and is proportional to outcome.
the size of the interval. Quantitative forecasts Forecasts based on a formal model,
Poisson probability distribution A description of the prob along with historical data for the variable of interest.
abilities associated with the possible values of a Poisson ran Quantitative variable A variable that assumes meaningful
dom variable. numerical values.
Poisson random variable The number of successes over a Quartiles Any of the three values that divide the ordered
given interval of time or space in a Poisson process. data into four equal parts, so that each part represents 1/4 of the
Polygon A graph of a frequency or relative frequency distri sample or population.
bution in which lines connect a series o f neighboring points, Randomized block design In ANOVA, the design of an
where each point represents the midpoint of a particular class experiment in which units within each block are randomly
and its associated frequency or relative frequency. assigned to each o f the treatments.
Polynomial regression models Regression models that de Random variable A function that summarizes outcomes of
scribe various numbers of sign changes in the influence of an an experiment with numerical values.
explanatory variable on a response variable. Range The difference between the maximum and the mini
Polynomial trend models Models used for time series mum values in a data set.
that reverse direction. A quadratic trend model is used for Ratio-scaled data Values of a quantitative variable that can
one change of direction, a cubic trend model is used for two be categorized and ranked, and in which differences between
changes of direction, etc. values are meaningful; in addition, a true zero point (origin)
Population The complete collection of items with the char exists.
acteristic we wish to understand; all items of interest in a sta Real return Investment return that is adjusted for the change
tistical problem. in purchasing power due to inflation.
Portfolio A collection of assets. Regression analysis A statistical method for analyzing the
Portfolio variance A measure reflecting a portfolio’s risk relationship between variables. The method assumes that one
that depends on the variance of the individual assets in the port variable, called the response variable, is influenced by other
folio and on the covariance between the assets. variables, called the explanatory variables.
Positively skewed (right-skewed) distribution A distribu Rejection region In a hypothesis test, a range of values such
tion in which extreme values are concentrated in the right tail that if the value of the test statistic falls into this range, then the
of the distribution. decision is to reject the null hypothesis.
Relative frequency distribution A frequency distribution Skewness coefficient A measure of how symmetric data are
that shows the fraction or proportion of observations in each about the mean. Symmetric data have a skewness coefficient
category of qualitative data or class of quantitative data. of zero.
Residual (e) In regression analysis, the difference between Smoothing techniques In time series analysis, methods to
the predicted and the actual value, that is, e = y − ŷ . provide forecasts if short-term fluctuations represent random
Residual plots In regression analysis, graphs o f residuals departures from the structure with no discernible systematic
against variables and time to check for model inadequacies. patterns.
Response variable In regression analysis, the variable that Spearman rank correlation test A nonparametric test to de
we assume is affected by other variables. It is also called the termine whether two variables are correlated.
dependent variable, the explained variable, the predicted vari Standard deviation The positive square root of the variance;
able, or the regressand. a common measure of dispersion.
Risk averse consumer Someone who may decline a risky Standard error An estimate o f the standard deviation of an
prospect even if it offers a positive expected gain. estimator.
Risk loving consumer Someone who may accept a risky Standard error of the estimate The standard deviation of
prospect even if the expected gain is negative. the residual, used as a goodness-of-fit measure for regression
Risk neutral consumer Someone who completely ignores analysis.
risk and always accepts a prospect that offers a positive ex Standard normal distribution A special case of the normal
pected gain. distribution with a mean equal to zero and a standard deviation
Runs test A procedure used to examine whether the elements (or variance) equal to one.
in a sequence appear in random order. Statistic A characteristic o f a sample used to make infer
Sample A subset of a population of interest. ences about an unknown population parameter; a statistic is a
random variable whose value depends on the chosen random
Sample space (S) A record o f all possible outcomes o f an
sample.
experiment.
Statistical Quality Control Statistical techniques used to
Sampling distribution The probability distribution of an
develop and maintain a firm’s ability to produce high-quality
estimator.
goods and services.
Scatterplot A graphical tool that helps in determining
Stem-and-leaf diagram A visual method o f displaying
whether or not two variables are related in some systematic
quantitative data where each value of a data set is separated
way. Each point in the diagram represents a pair of known or
into two parts: a stem, which consists o f the leftmost digits, and
observed values of the two variables.
a leaf, which consists of the last digit.
Seasonal component Repetitions of a time series over a one-
Stochastic relationship In regression analysis, a relationship
year period.
in which the explanatory variables do not exactly predict the
Seasonal dummy variables Dummy variables used to cap response variable.
ture the seasonal component from a time series.
Stratified random sampling A population is first divided
Seasonal index A measure of the seasonal variation within a up into mutually exclusive and collectively exhaustive groups,
time series; used to deseasonalize data. called strata. A stratified sample includes randomly selected
Selection bias A systematic exclusion of certain groups from observations from each stratum, which are proportional to the
consideration for a sample. stratum’s size.
Semi-log regression models Regression models in which Subjective probability A probability value based on per
some but not all variables are transformed into natural logs. sonal and subjective judgment.
Serial correlation In regression analysis, a situation where Symmetric distribution A distribution that is a mirror image
the error term is correlated across observations. of itself about its mean.
Sharpe ratio The ratio of reward to risk, with the reward Systematic patterns In time series, patterns caused by a set
specified in terms of the m ean and the risk specified in terms o f of identifiable components: the trend, seasonal, and the cyclical
the standard deviation. components.
Sign test A nonparametric test to determine whether signifi t Distribution A family o f distributions that are similar to
cant differences exist between two populations using matched- the z distribution except that they have broader tails. They are
pairs sampling with ordinal data. identified by their degrees of freedom that determine the extent
Simple linear regression model In regression analysis, one of broadness.
explanatory variable is used to explain the variability in the Test of independence A goodness-of-fit test analyzing the
response variable. relationship between two qualitative variables. Also called a
Simple price index For any item, the ratio of the price in a chi-square test of a contingency table.
given time period to the price in the base period, expressed as Test of individual significance In regression analysis, a test
a percentage. that determines whether an explanatory variable has an indi
Simple random sample A sample of n observations that has vidual statistical influence on the response variable.
the same probability o f being selected from the population as Test of joint significance In regression analysis, a test to
any other sample of n observations. determine whether the explanatory variables have a joint
statistical influence on the response variable; it is often re Unbiasedness An estimator is unbiased if its expected value
garded as a test of the overall usefulness of a regression model. equals the unknown population parameter being estimated.
Test of linear restrictions In regression analysis, a test to Unconditional probability See Marginal probability.
determine if the restrictions specified in the null hypothesis are Unsystematic patterns In time series, patterns caused by the
invalid. presence of an irregular or random error term.
Time series A set of sequential observations of a variable Unweighted aggregate price index An aggregate price
over time. index based entirely on aggregate prices with no emphasis
Total probability rule A rule that expresses the unconditional placed on quantity.
probability of an event, P (A ), in terms of probabilities condi Variable A general characteristic being observed on a set
tional on various mutually exclusive and exhaustive events. The of people, objects, or events, where each observation varies in
total probability rule conditional on two events B and Bc is kind or degree.
P (A ) = P(A ᑎ B ) + P (A ᑎ Bc) = P (A \B )P(B ) + P(A \B c)P(Bc). Variance The average of the squared differences from the
Tukey’s honestly significant differences (HSD) method In mean; a common measure of dispersion.
ANOVA, a test that determines which means significantly dif Weighted aggregate price index An aggregate price index
fer by comparing all pairwise differences of the means. that gives higher weight to the items sold in higher quantities.
Two-tailed hypothesis test A test in which we can reject the Wilcoxon rank-sum test A nonparametric test to determine
null hypothesis on either side of the hypothesized value of the whether two population medians differ under independent
population parameter. sampling. Also known as the Mann-W hitney test.
Two-way ANOVA A statistical technique used to examine si Wilcoxon signed-rank test A nonparametric test to deter
multaneously the effect of two factors on the mean. These tests mine whether a sample could have been drawn from a popula
can be conducted with or without the interaction of the factors. tion having a hypothesized value as its median; this test can
Type I error In a hypothesis test, this error occurs when the also be used to determine whether the median difference differs
decision is to reject the null hypothesis, but the null hypothesis from zero under matched-pairs sampling.
is actually true. Within-treatments variability In ANOVA, a measure of the
Type II error In a hypothesis test, this error occurs when variability within each sample.
the decision is to not reject the null hypothesis, but the null bar chart A control chart that monitors the central ten
hypothesis is actually false. dency o f a production process.
Unbalanced data A completely randomized ANOVA design z table A table providing cum ulative probabilities for
where the number of observations are not the same for each positive or negative values o f the standard normal random
sample. variable Z .
p. 3, © Michael Schmitt/Workbook Stock/ p. 207, © M elanie Conner/Getty Images p. 423, © Digital Vision/Getty Images
CREDITS
Getty Images p. 223, © Joe Raedle/Getty Images p. 447, © AP Photo/David J. Phillip
p. 7, © Digital Vision/PunchStock p. 233, © Ryan McVay/Photodisc/Getty p. 449, © Kick Images/Photodisc/Getty
p. 9, © Ed Young/Agstockusa/age fotostock, Images Images
America, Inc. p. 241, © Car Culture/Corbis p. 457, © M artin Thomas Photography/
p. 13, © Dennis Welsh/UpperCut Im ages/ p. 262, © The McGraw-Hill Companies, Inc./ Alamy
Getty Images M ark A. Dierker, photographer p. 464, © Scott Boehm/Getty Images
p. 17, © A. Ramey/PhotoEdit Inc. p. 263 right, © The McGraw-Hill Com pa p. 483, © Elsa/Getty Images
p. 18, sbk_20d pictures/Flickr/Getty Images nies, Inc./Andrew Resek, photographer p. 491, © N ick Dolding/Photographer’s
p. 37, © Brand X Pictures p. 263 left, © AP Photo/Paul Sakuma Choice RF/Getty Images
p. 45, © Rubberball/Getty Images p. 271, © Fotosearch/age fotostock, America, p. 508, © Richard T. Nowitz/Corbis
p. 53, © Jose Luis Pelaez, Inc./Getty Images Inc. p. 511, © Photodisc/Getty Images
p. 58, © Sebastian Pfuetze/Taxi/Getty Images p. 291, © A sia Images Group/Getty Images p. 517, © Jose Luis Pelaez, Inc./Blend
p. 298, © Ariel Skelley/Blend Images LLC Images/Corbis
PHOTO
p. 73, © Ingram Publishing
p. 87, © Mike Watson Images/PunchStock p. 305, © Robin Nelson/PhotoEdit, Inc. p. 528, © Tom Stewart/Corbis
p. 95, © Smith Collection/Iconica/Getty p. 320, © Chris Hondros/Getty Images p. 537, © AP Photo/M ark J . Terrill, Pool
Images p. 328, © Ingram Publishing p. 545, © Grace Davies/age fotostock,
p. 102, © AP Photo/Gene J. Puskar p. 335, © Steve Cole/Vetta/Getty Images America, Inc.
p. 116, © Digital Vision/Masterfile p. 352, © Spencer Grant/PhotoEdit, Inc. p. 569, © David Paul Morris/Bloomberg via
p. 128, © R olf Bruderer/Blend Images/Getty p. 354, © A P Photo/Gerry Broome Getty Images
Images p. 361, © Liu Jin/AFP/Getty Images p. 575, © AP Photo/Evan Vucci
p. 135, © AP Photo/Kevin P. Casey p. 372, © Jean Baptiste Lacroix/W ireImage/ p. 585, © Jeff Greenberg/Alamy
p. 160, © Bubbles Photolibrary/Alamy Getty Images p. 596, © Peter Titmuss/Alamy
p. 166, © Image Source/Alamy p. 379, © JGI/Blend Images/Getty Images p. 603, © Bettmann/Corbis
p. 173, © QxQ IMAGES/Datacraft/Getty p. 387, © Index Stock/Photolibrary/age p. 609, © Tim Brown/Stone/Getty Images
Images fotostock, America, Inc. p. 628, © Steve Cole/Getty Images
p. 191, © Mewes, K./Healthy Food Images/ p. 400, © Jeff Greenberg/PhotoEdit Inc. p. 639, © Studio Eye/Corbis
age fotostock, America, Inc. p. 415, © Mitchell Funk/Photographer’s
p. 201, © Im age Source/Getty Images Choice/Getty Images
A Binomial probability distributions
INDEX
Acceptance sampling, 227 constructing, 150– 152
Addition rule, 105– 107 defined, 150
Adidas, 19– 2 1 , 66, 95, 114, 116, 3 6 1 , 372 with Excel, 154– 156
Adjusted close prices, 587– 588 formula, 152
Adjusted coefficient of determination, 446, 560 sampling distribution of sample proportion, 219
Adjusted seasonal index, 564– 565 Binomial random variables
Aggregate price indices, 592 defined, 150
unweighted, 592– 593 expected values, 152– 153
weighted, 593–596 standard deviation, 152– 153
Alpha, 461 variance, 152– 153
Alstead, Troy, 4 Blocked outcomes, 404
Alternative hypothesis, 272– 273 BLS; see Bureau of Labor Statistics
Analysis of variance (ANOVA); see also One-way ANOVA; BLUE; see Best linear unbiased estimator
Two-way ANOVA Bonds; see Returns
ANOVA table, 392 Boston Celtics, 537
balanced designs, 397– 398 Box plots, 62–63
completely randomized design, 388 Brown, Scott, 4, 211– 212
defined, 388 Bryant, Kobe, 537
unbalanced designs, 397– 398 Bureau of Economic Analysis (BEA), 7
uses of, 386, 388, 390 Bureau of Labor Statistics (BLS), 6, 7, 599
Annualized returns, 65 Business cycles, 562
ANOVA; see Analysis of variance; One-way ANOVA;
Two-way ANOVA c
ANOVA table, 392 Capital asset pricing model (CAPM), 461 –462
Anti-log function, 500 Capital gains yields, 586, 587
A priori probabilities, 101, 102 CAPM; see Capital asset pricing model
Arithmetic mean, 54– 55, 65–66 Causal forecasting models, 546, 572– 573
Arizona Cardinals, 102 Causation, correlation and, 5, 428
Under Armour Inc., 95, 114– 117, 361, 372 Census Bureau, U.S., 7, 56
Asset returns; see Returns Centerline, 228
Assets; see Portfolios Central limit theorem (CLT)
Assignable variation, 227 sampling distribution o f sample mean, 216– 217
Associated Press, 298 sampling distribution o f sample proportion, 220–221, 230
Asymptotic distributions, 178 Central location measures
Autocorrelation, 476 arithmetic mean, 54–55, 65–66
Autoregressive models, 572 defined, 54
Average growth rates, 66–67 Excel calculations, 58– 59
Averages; see Mean geometric mean, 54, 64–66
median, 56–57
mode, 57–58
B weighted mean, 8 1
Balanced data, 397–398 Chance variation, 227
Bar charts, 21–22, 23 Charts; see Control charts; Graphical displays
Bayes, Thomas, 121 Chebyshev, Pavroty, 77
B ayes’ theorem, 118, 121– 123 Chebyshev’s theorem, 76–77, 78
BEA; see Bureau of Economic Analysis Chi-square (χ 2) distribution
Bell curve, 178; see also Symmetric distributions characteristics, 336–337
Bell Telephone Laboratories, 227 defined, 336
Bernoulli, James, 150 degrees o f freedom, 336
Bernoulli process, 149– 150, 362 locating values and probabilities, 337– 338
Best linear unbiased estimator (BLUE), 476 Chi-square (χ 2) table, 337– 338, 650– 651
Beta, 461–462 Chi-square tests
Between-treatments estimate o f population variance, 390 goodness-of-fit
Between-treatments variability, 390 for independence, 368–372
Bias, 2 0 8 , 211 for multinomial experiments, 363– 367
Binary choice models for normality, 374– 376
defined, 530 Jarque-Bera, 376– 377
linear probability model, 530–532, 533 Classes, 25–27
logit model, 532– 535 Clinton, Hillary, 211
Binomial experiments, 149– 150, 362 CLT; see Central limit theorem
Cluster sampling, 211– 212 Correlation coefficient
Coakley, Martha, 4, 211– 212 calculating, 85
Coefficient of determination, 444– 146, 507–508 defined, 84
Coefficient of variation (CV), 71– 72 hypothesis tests, 426–427
Combination formula, 126 population, 84, 426–4 2 7
Complement, of event, 98, 105 o f portfolio returns, 147– 148
Complement rule, 105 sample, 84, 425
Completely randomized designs, 388 Correlation-to-causation fallacy, 5
Conditional probability, 108– 109 Counting rules
Confidence intervals combination formula, 126
defined, 245 factorial formula, 125– 126
for differences in means, 306– 308 permutation formula, 126– 127
individual significance tests, 460–461 Covariance
interpreting, 246–247 formula, 84– 85
margin of error, 245–246, 259– 260 population, 83
for mean difference, 316 of portfolio returns, 147– 148
of population mean sample, 83, 424
with known standard deviation, 245– 249, 259 CPI; see Consumer Price Index
with unknown standard deviation, 253–254, 259 Critical values
of population proportion, 257– 258, 260 defined, 282
of population variance, 339–340 hypothesis testing, 281– 284, 289
for population variance ratio, 347– 348 Cross-sectional data, 6, 7
predicted values, 472–474 Cubic model, 497–498
for proportion differences, 323– 324 Cubic trend models, 560
sample sizes and, 259– 261 Cumulative distribution function, 137, 174
two-tailed hypothesis tests, 285 Cumulative frequency distributions
width and precision, 248– 249 defined, 28
Consistency, of point estimators, 244 ogives, 34– 36
Constants, 213 Cumulative probability distributions, 138
Consumer Price Index (CPI), 592, 599–601, 603; see also Cumulative relative frequency distributions, 29
Inflation rates; Price indices Curvilinear relationships, 42; see also Polynomial
Consumers, risk and, 143– 144 regression models
Contingency tables CV; see Coefficient o f variation
chi-square tests, 368– 372 Cyclical patterns, 548, 562
defined, 113, 368
using, 114– 115 D
Continuous probability distributions; see also Data; see also Time series data; Variables
Normal distribution cross-sectional, 6, 7
exponential, 194– 196 measurement scales, 9– 12
lognormal, 196– 199 sources, 7
uniform, 175– 176 types, 6–7
Continuous random variables, 9, 136, 137, 174 websites, 7
Continuous uniform distribution, 175– 176 Deciles, 62
Continuous variables, 9 Decision rules
Control charts critical value approach, 283
defined, 227 p -value approach, 279
p -bar charts, 227, 230–231 Decomposition analysis, 562– 563
for qualitative data, 229–231 extracting seasonality, 563– 565
for quantitative data, 227–229 extracting trend, 565– 566
x-bar charts, 227– 229 forecasting, 566– 567
Control variables; see Explanatory variables Deflated time series, 598–600
Correlation Degrees of freedom (d f), 251, 252
auto-, 476 Dependent events, 110
causation and, 5, 428 D ependent variable; see Response variable
serial, 476, 480–4 8 1 Descriptive statistics, 5
Spearman rank correlation test, 625– 628 Detection approach, 227
spurious, 5, 428 Deterministic relationships, 430
Correlation analysis d f; see Degrees of freedom
hypothesis tests, 426–427 Discrete choice models; see Binary choice
limitations of, 427–4 2 8 models
Discrete probability distributions Estimation; see Confidence intervals; Maximum likelihood
binomial estimation; Point estimators
constructing, 150– 152 Estimators, 213
defined, 150 Events
with Excel, 154– 156 complement of, 98, 105
formula, 152 defined, 96–97
sampling distribution of sample proportion, 219 dependent, 110
defined, 137 exhaustive, 97
formula, 138 independent, 109, 110, 111
graphical displays, 138 intersection of, 97–98
hypergeometric, 162– 165 mutually exclusive, 97, 107
Poisson, 157– 161 probabilities, 96
Discrete random variables union of, 97, 98,1 0 5 – 106
defined, 9, 136 Excel
expected values, 141– 142 ANOVA problems, 392– 393
experiments, 136– 137 bar charts, 21–22
standard deviation, 142– 143 binomial probabilities, 154– 156
variance, 142– 143 central location measures, 58– 59
Discrete uniform distributions, 139 chi-square tests, 365– 367
Discrete variables, 9 confidence intervals, 254–255
Dispersion measures, 68 control charts, 229, 231
coefficient of variation, 71–72 correlation coefficient, 85, 426
Excel calculations, 72 covariance, 85, 426
mean absolute deviation, 69– 70 dispersion measures, 72
range, 69 exponential distribution, 195– 196
standard deviation, 70– 71, 76– 78 exponential smoothing, 553
variance, 70– 71 histograms, 31– 33
Distribution–free tests; see Nonparametric tests hypergeometric probabilities, 164– 165
DJIA; see Dow Jones Industrial Average hypothesis testing, 285–286
Domino’s Pizza, 632 hypothesis tests for differences in means, 310– 312
Dow Jones Industrial Average (DJIA), 9– 10 hypothesis tests for mean difference, 318– 320
Dummy variable trap, 521 lognormal distribution, 198– 199
Dummy variables moving average, 553
defined, 518 multiple regression, 438–439
interaction variables and, 525– 528 normal distribution, 191– 192
for multiple categories, 521–522 ogives, 36
seasonal, 562, 567–569 pie charts, 21
significance tests, 520–521 Poisson probabilities, 160– 161
use of, 518–520 polygons, 34–35
Dunkin’ Donuts, 207 p -values, 290–291, 341– 342, 350– 351
Duracel, 6 regression statistics, 443
residual plots, 477
E sample regression equation, 434– 136
eBay, 134 scatterplots, 43, 432
Economist, 7 smoothing techniques, 553
Efficiency, of point estimators, 243 standard error of the estimate, 443
Empirical probabilities, 100– 101, 102 trendlines, 432
Empirical rule, 77– 78, 184– 186 two-way ANOVA
Endogeneity, 476, 481 with interaction, 412–413
Environmental Protection Agency, 241 no interaction, 406–407
Error sum of squares (SSE) Exhaustive events, 97
ordinary least squares, 432 Expected frequencies, 369– 370
in two-way ANOVA, 405–406, 412 Expected portfolio returns, 146– 147
within-treatments estimate of population variance, 391 Expected values, 141– 142
Errors; see also Margin of error; Standard error Experiments; see also Events
random, 548 binomial, 149– 150, 362
Type I, 275– 276 defined, 96
Type II, 275–276 multinomial, 362– 363
espn.com, 7 outcomes, 96
Estimates, 213 Explained variation, 444
Explanatory variables; see also Dummy variables ogives, 34– 36
defined, 430 polygons, 33–35
endogeneity, 481 qualitative data, 18– 19
multicollinearity, 477–478 quantitative data, 25– 29
in multiple regression, 437–438 relative, 18– 19, 28– 29
qualitative, 518
in simple linear regression, 430 G
Exponential function, 500 Gallup, 5
Exponential model, 502, 503– 504 Geometric mean, 54, 64– 66
Exponential probability distribution, 194– 196 Geometric mean returns, 64– 66
Exponential smoothing, 550– 553 Goodness-of-fit tests
Exponential trend models adjusted coefficient o f determination, 446
estimating, 556, 558– 559 coefficient of determination, 444–4 4 6
formula, 557 for independence, 368– 372
seasonal dummy variables, 567– 568 for multinomial experiments, 363– 367
uses of, 556, 557–558 for normality, 374– 376
for regression analysis, 441
standard error of the estimate, 442– 444
F Google, 7
F distribution Gossett, William S., 251n
characteristics, 344– 345 Grand mean, 390
locating values and probabilities, 345– 347 Graphical displays
F table, 345– 347, 652– 655 bar charts, 21– 22, 23
F test box plots, 62–63
assumptions, 610 guidelines, 22– 23
one-tailed, 463– 164 histograms, 30– 33
partial, 467–470 ogives, 34– 36
Factorial formula, 125– 126 pie charts, 20–21
Federal Reserve, 94 polygons, 33– 35
Federal Reserve Economic Data (FRED), 7 residual plots, 476–477
Fidelity Investments, 53, 73, 335, 352, 609, 628 scatterplots, 41–43, 424, 432
Finite population correction factor, 224– 225 stem-and-leaf diagrams, 39– 4 1
Fisher, Irving, 588 Grouped data, summarizing, 79– 81
Fisher equation, 588– 589 Growth rates, average, 66–67
Fisher’s least significant difference (LSD) method, Guinness Brewery, 2 5 1n
395– 397, 398, 408
Fitted value, 547 H
Forecasting; see also Smoothing techniques Hartley, H. O., 657
causal models, 546, 572– 573 Heteroskedasticity, 478–479
decomposition analysis, 562– 567 Histograms
model selection, 547 constructing, 30, 31– 33
noncausal models, 546 defined, 30
qualitative, 546 relative frequency, 30
quantitative, 546 shapes, 31
trend models Home Depot, 263–264
exponential, 556– 559, 567–568 Homoskedasticity, 476, 478
linear, 555– 556, 567 Hoover, Herbert, 208
polynomial, 559– 560 HSD; see Tukey’s honestly significant differences method
quadratic, 559– 560 Hypergeometric probability distributions, 162– 165
Fortune, 7 Hypergeometric random variables, 163
Fraction defective charts, 230 Hypotheses
FRED; see Federal Reserve Economic Data alternative, 272–273
Frequencies, expected, 369– 370 null, 272– 273, 275– 276
Frequency distributions Hypothesis tests; see also Analysis o f variance;
classes, 25– 27 Nonparametric tests
constructing, 26– 27 for correlation coefficient, 426–4 2 7
cumulative, 28 for differences in means, 308– 312
defined, 28 of individual significance, 458– 461
graphical displays interpreting results, 287
charts, 20–22 of joint significance, 462–4 6 4
histograms, 30– 33 left-tailed, 273
for mean difference, 317– 320 J
one-tailed, 273–275 Jarque-Bera test, 376– 377
of population mean, with known standard J.D. Power and Associates, 230
deviation, 277– 278 Johnson & Johnson, 461
critical value approach, 281– 284 Joint probability, 110– 111
p -value approach, 278– 281, 282 Joint probability tables, 115– 116
of population mean, with unknown standard deviation, Joint significance, tests of, 462–464
288– 291
of population median, 611 K
of population proportion, 294– 296 Kennedy, Edward, 4, 211
for population variance, 340– 341 Kruskal–Wallis test, 621–624
for population variance ratio, 348– 350 Kurtosis, 59
power, 276 Kurtosis coefficients, 376
for proportion differences, 324– 326
rejecting null hypothesis, 272, 273 L
right-tailed, 273 Lagged regression models, 572–573
significance level, 279 Landon, Alf, 208, 209
two-tailed, 273–274, 285 Laspeyres price index, 593– 594, 595– 596
Type I and Type II errors, 275–276 LCL; see Lower control limit
Left-tailed hypothesis tests, 273
I Linear probability model (LPM), 530–532, 533
IBM, 6 Linear relationships; see also Multiple regression model;
Income yields, 586 Simple linear regression model
Independence, chi-square test for, 368– 372 correlation coefficient, 84, 85, 425, 427–428
Independent events, 109, 110, 111 covariance, 83, 84– 85, 424
Independent random samples multicollinearity, 477–478
defined, 306 negative, 83
Wilcoxon rank-sum test, 614, 616–619 positive, 83
Independent variables; see Explanatory variables scatterplots, 42, 424, 432
Index numbers; see also Indices slopes, 431
base periods, 591– 592 Linear restrictions, general test of, 467–470
defined, 590 Linear trend models
Indicator variables; see Dummy variables forecasts, 555– 556
Indices; see also Price indices formula, 555
base periods, 591–592 seasonal dummy variables, 567
seasonal Literary Digest poll, 208– 209
adjusted, 564–565 Logarithmic model, 502– 503, 504– 507
unadjusted, 564 Logarithms
Individual significance, tests of, 458–461 log-log model, 500– 502, 504–507
Inferential statistics, 5 natural, 500
Inflation; see also Price indices semi-log model, 502
effects, 598 Logit model, 532–535
returns and, 588–589 Log-log model, 500– 502, 504–507
Inflation rates Lognormal distribution, 196– 199
calculating, 600–601 Los Angeles Lakers, 537
defined, 600 Lower control limit (LCL), 228–229, 230
expected, 588 Lowe’s, 263–264
In-sample criteria, 547 LPM; see Linear probability model
Instrumental variables, 481 LSD; see Fisher’s least significant difference method
Interaction variables, 525– 528
Intercept dummy, 519 M
Interquartile range (IQR), 63, 69 MAD; see Mean absolute deviation
Intersection, of events, 97– 98 Main effects, 413
Interval estimates, 472–474; see also Confidence intervals Major League Baseball (MLB), 457, 464–465, 483–484
Interval scale, 12 M ann-Whitney test; see Wilcoxon rank-sum test
Inverse transformation, 189– 191 Margin of error, 245–246, 259– 260
Investment returns; see also Returns Marginal probability, 108
calculating, 586– 587 Matched outcomes, 404
income component, 586 Matched-pairs sampling
price change component, 586, 587 mean difference, 315, 316–319
IQR; see Interquartile range ordinal data, 631–633
M atched-pairs sampling (Continued) response variable, 437– 438
recognizing, 316 sample regression equation, 438– 439
sign test, 631–633 test o f linear restrictions, 467–4 7 0
Wilcoxon signed-rank test, 320, 615– 616 tests of individual significance, 458–461
Matthews, Robert, 428 tests of joint significance, 462–464
M aximum likelihood estimation (MLE), 532 Multiplication rule, 110– 111
M cD onald’s, 194, 207 Mutually exclusive events, 97, 107
M ean; see also Population mean; Sample mean
arithmetic, 54– 55, 65– 66
N
for frequency distribution, 80– 81
Nasdaq, 9– 10
geometric, 54, 64– 66
National Basketball Association (NBA), 537
moving average, 548– 550
Natural logarithm, 500
weighted, 81
NBA; see National Basketball Association
Mean absolute deviation (MAD), 69– 70, 547
Negatively skewed distributions, 31, 59
Mean difference
New York Stock Exchange (NYSE), 9– 10, 546
confidence intervals, 316
New York Times, 7
hypothesis test, 317– 320
Nike, Inc., 95, 114, 115, 116, 3 61,372, 545, 563, 565,
matched-pairs experiments, 315, 316– 319
566– 567, 568, 569
Mean square error (MSE), 391. 412, 442, 547
Nominal returns, 588; see also Returns
Mean square for factor A (MSA), 405, 411
Nominal scale, 9– 10, 12, 18
Mean square for factor B (MSB), 405, 411
Nominal terms, 598
Mean square for interaction (MSAB), 412
Noncausal models, 546
M ean square for treatments (MSTR), 390
Nonlinear models; see Polynomial regression models
M ean-variance analysis, 74– 75
Nonparametric tests
Measurement scales, 9
compared to parametric tests, 608, 610, 628
interval, 12
Kruskal-Wallis test, 621– 624
nominal, 9– 10, 12, 18
sign test, 631– 633
ordinal, 10, 12, 18
Spearman rank correlation test, 625–628
ratio, 12
uses of, 610
Median; see also Population median
W ilcoxon rank-sum test, 312, 614–615, 616– 619
calculating, 56– 57
W ilcoxon signed-rank test
defined, 56
for matched-pairs sample, 615– 616
runs above and below, 635, 636–637
for population median, 610– 613
M ethod of least squares; see Ordinary least squares
software, 616
M ethod of runs above and below median, 635,
uses of, 320
636–637
N onresponse bias, 209
Michigan State University, 603
Normal curve, 178– 179
Microsoft Corporation, 586, 588
Normal distribution
Minitab
characteristics, 178
Kruskal-Wallis test, 623
defined, 177
runs test, 637
empirical rule, 77– 78, 184– 186
W ilcoxon rank-sum test, 619
of error term, 476
W ilcoxon signed-rank test, 616
probability density function, 178– 179
MLB; see M ajor League Baseball
standard, 179– 180
MLE; see Maximum likelihood estimation
using Excel, 191– 192
Mode, 57– 58
Normal transformation, 187– 189
Model selection, for forecasting, 547
Normality
M oving average methods, 548– 550, 553
goodness-of-fit test, 374– 376
m -period moving average, 548– 550
Jarque-Bera test, 376– 377
M SA ; see Mean square for factor A
Null hypothesis; see also Hypothesis tests
MSAB; see Mean square for interaction
defined, 272
MSB; see M ean square for factor B
formulating, 272– 273
MSE; see Mean square error
rejecting, 272, 273
MSTR; see Mean square for treatments
Type I and Type II errors, 275– 276
Multicollinearity, 476, 477– 478
NYSE; see New York Stock Exchange
Multinomial experiments, 362– 363
Multiple R, 445– 446
Multiple regression model; see also Regression analysis o
defined, 438 Obama, Barack, 5, 211
explanatory variables, 437–438 Objective probabilities, 102
interval estimates for predictions, 472–474 Odds ratios, 102– 103
Ogives, 34–36 Polynomial trend models, 559–560
Ohio State University, 5 Pooled estimates, 307
Olds, E. G ., 658 Population coefficient of variation, 71
OLS; see Ordinary least squares Population correlation coefficient, 84, 426– 427
One-tailed hypothesis tests, 273– 275 Population covariance, 83
One-way ANOVA Population mean
between-treatments estimate, 390 confidence intervals
defined, 388 of differences, 306– 308
example, 389– 390 with known standard deviation, 245– 249, 259
samples, 388 with unknown standard deviation, 253–254, 259
test statistic, 391– 393 differences between
within-treatments estimate, 390, 391 confidence intervals, 306– 308
Ordinal data, matched-pairs sampling, 631–633 Fisher’s least significant difference method, 395–397,
Ordinal scale, 10, 12, 18 398, 408
Ordinary least squares (OLS) hypothesis test, 308–312
as best linear unbiased estimator (BLUE), 476 test statistic, 309– 310
lines fit by, 432– 433 Tukey’s honestly significant differences method,
normal distribution, 458 397– 399, 408
properties, 476 expected values, 141– 142
Outliers, 55 formula, 54–55
Out-of-sample criteria, 547 matched-pairs experiments, 315– 316
Population mean absolute deviation, 69
P Population median
Paasche price index, 594– 596 Kruskal-Wallis test, 621– 624
Parameters; see Population parameters matched-pairs experiments, 615– 616
Parametric tests; see also F test; t test Wilcoxon rank-sum test, 614–615,
assumptions, 610 616– 619
compared to nonparametric tests, 608, 610, 628 Wilcoxon signed-rank test
Partial F test, 467–470 for matched-pairs sample, 615– 616
p -bar charts, 227, 230–231 for population median, 610–613
Pearson, E. S., 657 software, 616
Pearson correlation coefficient; see Correlation coefficient Population parameters
Percent defective charts, 230 constants, 213
Percent frequencies, 19 defined, 5
Percentiles inferences about, 6, 208, 213
calculating, 61–62 mean as, 55
defined, 61 Population proportion
quartiles, 62–63 confidence intervals, 257– 258, 260
Permutation formula, 126– 127 differences between
Pie charts, 20– 21 applications, 322– 323
Pittsburgh Steelers, 102 confidence intervals, 323– 324
Point estimates, 242 hypothesis test, 324– 326
Point estimators test statistic, 325
consistent, 244 hypothesis tests, 294– 296
defined, 242 Population standard deviation, 70
efficient, 243 Population variance
properties, 242– 244 between-treatments estimate, 390
unbiased, 242– 243 confidence intervals, 339– 340
Poisson, Simeon, 157 formula, 70
Poisson probability distribution, 157– 161 hypothesis test, 340– 341
Poisson process, 157– 158 quality control studies, 336
Poisson random variables, 157, 158– 159 ratio between
Polls, 5, 208– 209, 210–212 confidence intervals, 347– 348
Polygons, 33– 35 estimating, 344– 345
Polynomial regression models hypothesis test, 348– 350
cubic, 497– 498 test statistic, 349
defined, 497 test statistic, 340
quadratic within-treatments estimate, 390, 391
compared to linear model, 492–497 Populations
formula, 492 defined, 5
predictions, 492 samples and, 6, 208
Portfolio returns; see also Returns Proportion; see Population proportion; Sampling distribution
correlation coefficient, 147– 148 o f sample proportion
covariance, 147– 148 Purchasing power; see Inflation; Price indices
expected, 146– 147 p -values
defined, 279
standard deviation, 148– 149
hypothesis testing, 278– 281, 282, 341– 342
variance, 146– 148
Portfolio risk, 147
Portfolio weights, 147 Q
Portfolios, defined, 146 Quadratic regression model
compared to linear model, 492–497
Positively skewed distributions, 31, 59
formula, 492
Posterior probabilities, 121– 122
predictions, 492
Power o f the test, 276 Quadratic trend models, 559– 560
PPI; see Producer Price Index Qualitative forecasting, 546; see also Forecasting
Predicted regression equation, 431 Qualitative response models; see Binary choice models
Predicted value, 547 Qualitative variables; see also Binary choice models; Dummy
Prediction intervals, 472, 474 variables; Population proportion
Predictor variables; see Explanatory variables contingency tables, 113– 115
Price indices; see also Inflation rates control charts, 229– 231
defined, 8
aggregate, 592
frequency distributions, 18– 19
unweighted, 592– 593 graphical displays, 20–22
weighted, 593–596 nominal, 9– 10, 12, 18
base periods, 591– 592 ordinal, 10, 12, 18
base values, 590 in regression, 518
Consumer Price Index, 592, 599–601, 603 summarizing, 18
deflated time series, 598– 600 Quality control; see Statistical quality control
Laspeyres, 593– 594, 595–596 Quantitative forecasting, 546; see also Forecasting
Paasche, 594– 596 Quantitative variables
control charts, 227– 229
Producer Price Index, 592, 599–600
defined, 8–9
simple, 590– 592
frequency distributions, 25–29
Prices
in regression, 518
adjusted close, 587– 588
summarizing, 25– 26
capital gains yields, 586, 587
Quartiles, 62–63
Prior probabilities, 121– 122
Probabilities Quintiles, 62
a priori, 101, 102
assigning, 99 R
concepts, 96– 98 Random (irregular) error, 548
conditional, 108– 109
Random samples
defined, 96
for ANOVA, 388
empirical, 100– 101, 102
joint, 110– 111 completely randomized, 388
marginal, 108 independent
objective, 102 defined, 306
odds ratios, 102– 103 Wilcoxon rank-sum test, 614, 616– 619
posterior, 121– 122 simple, 209–210
prior, 121– 122 stratified, 210– 211
properties, 99 Random variables
subjective, 99, 102
binomial, 150, 152– 153
unconditional, 108, 119
continuous, 9, 136, 137, 174
Probability density function, 137, 174
Probability distributions; see also Continuous probability defined, 136
distributions; Discrete probability distributions discrete, 9, 136– 137
cumulative, 138 exponential, 194– 195
properties, 138 hypergeometric, 163
Probability mass function, 137 lognormal, 196
Probability rules normal transformation, 187– 189
addition rule, 105– 107 Poisson, 157, 158– 159
complement rule, 105
Randomized block designs, 404–406
multiplication rule, 110– 111
Ranges
total, 118– 121, 123
Probability tables, 119, 121 defined, 69
Probability trees, 119– 121, 150, 151 interquartile, 63, 69
Producer Price Index (PPI), 592, 599–600; see also Price Ratio scale, 12
indices Ratio-to-moving-average, 564
Reagan, Ronald, 603, 604
Real returns, 588–589
s
Sample coefficient of variation, 71
Real terras, 598 Sample correlation coefficient, 84, 425
Regression analysis, 430; see also Multiple regression model; Sample covariance, 83, 424
Simple linear regression model Sample mean; see also Sampling distribution of
assumptions, 475–477 sample mean
comparing models, 507–508 differences between, 306
goodness-of-fit tests, 441 formula, 54– 55
qualitative variables, 518 weighted, 81
quantitative variables, 518 Sample mean absolute deviation, 69
reporting results, 464 Sample proportion, 323; see also Sampling distribution of
violations o f assumptions, 477–481 sample proportion
Regression sum o f squares (SSR), 444 Sample regression equation
Rejection region, 281–282 multiple regression, 438–439
Relative frequency distributions simple linear regression, 431, 433–436
cumulative, 29 Sample spaces, 96; see also Events
qualitative data, 18– 19 Sample standard deviation, 70– 71
quantitative data, 28–29 Sample statistics
summarizing, 81 defined, 5
Residual plots, 476–4 77 estimators, 213
Residual sum of squares; see Error sum of squares mean as, 55
Residuals point estimators, 242
in-sample forecasts, 547 random, 213
in simple linear regression, 431 use of, 6, 208
Response variable Sample variance
binary choice models, 530 formula, 70– 71
defined, 430 properties, 336
expected values, 430–431 sampling distribution of, 336–337
in multiple regression, 437–438 sampling distribution of ratio of, 344–347
in simple linear regression, 430–431 Samples
Restricted models, 468 defined, 5
Returns independent random, 306
adjusted close prices and, 587–588 matched-pairs
annualized, 65 mean difference, 315, 316– 319
calculating, 586–587 ordinal data, 631– 633
excess, 75 recognizing, 316
geometric mean, 64–66 Wilcoxon signed–rank test, 320, 615– 616
historical, 587 populations and, 6, 208
mean-variance analysis, 74–75 sizes, 216, 259– 261
nominal, 588 Sampling
portfolio, 146– 149 acceptance, 227
real, 588–589 bias in, 208– 209, 211
risk and, 74– 75, 146, 147 cluster, 211– 212
risk-adjusted, 461–462 need for, 6, 208
Sharpe ratio, 75 simple random samples, 209– 210
Right-tailed hypothesis tests, 273 stratified, 210– 211
Risk with and without replacement, 163
measures, 336 Sampling distribution of sample mean
portfolio, 147 central limit theorem, 216– 217
returns and, 74–75, 146, 147 defined, 213
Risk aversion, 143– 144 expected values, 214, 239
Risk loving consumers, 144 hypothesis testing, 278
Risk neutral consumers, 144 normal distribution, 215– 216
Risk-adjusted returns, 461–462 standard deviation, 214
Roosevelt, Franklin D., 208, 209 variance, 214, 239
Rules of probability; see Probability rules Sampling distribution of sample proportion
Runs
central limit theorem, 220–221, 230
above and below median, 635, 636– 637 defined, 219
defined, 634 expected values, 219– 220, 239
Wald-Wolfowitz runs test, 634– 637 finite population correction factor, 225
Sampling distribution of sample proportion (Continued) interpreting, 76
standard deviation, 219– 220 population, 70
variance, 219, 239 of portfolio returns, 148– 149
Sampling distribution of sample variance, 336– 337 sample, 70–71
Scatterplots, 4 1–43, 424, 432 Standard error, of sample proportion, 219
Seasonal dummy variables, 562, 567– 569 Standard error of the estimate, 442–444
Seasonal indices Standard normal distribution
adjusted, 564– 565 areas under curve, 646– 647
unadjusted, 564 defined, 179– 180
Seasonal patterns, 548, 562 inverse transformation, 189– 191
Seasonality, extracting, 563–565 transformation, 187– 189
Seasonally adjusted series, 565– 566 z values, 179– 184
Selection bias, 209, 211 Standard normal table, 179– 180
Semi-log model, 502 Starbucks Corp., 4, 135, 159, 160, 207, 217, 221, 223, 305,
Serial correlation, 476, 480–481 317– 318,319, 320
Seton Hall University, 517, 528 Statistical quality control; see also Control charts
Sharpe, William, 75 acceptance sampling, 227
Sharpe ratio, 75 defined, 226– 227
Shewhart, Walter A., 227 detection approach, 227
Sign test, 631–633 population variance used in, 336
Significance levels, 279 sources o f variation, 227
Significance tests Statistical software; see Excel; Minitab
o f dummies, 520–521, 526 Statistics; see also Sample statistics
o f individual significance, 458– 461 descriptive, 5
o f interaction variables, 526– 528 importance, 4– 5
o f joint significance, 462–4 6 4 inferential, 5
Simple linear regression model; see also Regression analysis Stem-and-leaf diagrams, 39–41
assumptions, 430–43 1, 475–477 Stochastic relationships, 430
defined, 431 Stocks; see also Returns
estimated coefficients, 433 adjusted close prices, 587–588
explanatory variables, 430 dividend yields, 587
goodness-of-fit tests, 442– 444 Dow Jones Industrial Average, 9– 10
response variable, 430–431 Stratified random samples, 210–211
sample regression equation, 431, 433– 436 Studentized range table, 398, 656– 657
scatterplots, 432 Student’s t distribution; see t distribution
Simple price indices, 590–592 Subjective probabilities, 99, 102
Simple random samples, 209– 210 Sum o f squares due to treatments (SSTR), 390
Skewed distributions, 31 Sum of squares for factor A (SSA), 404–4 0 5 , 411
Skewness coefficients, 59, 376 Sum of squares for factor B (SSB), 404, 405, 411
Smoothing techniques Sum of squares for interaction of factor A and factor B (SSAB),
exponential, 550– 553 411– 412
moving average methods, 548– 550, 553 Surveys, 5, 208–209, 210– 212
uses of, 548 Symmetric distributions, 31, 77, 177, 178; see also Normal
Software; see Excel; Minitab distribution
Spearman rank correlation test, 625–628, 658 Systematic patterns, 548, 562
Spurious correlation, 5, 428
SSA; see Sum of squares for factor A T
SSAB; see Sum of squares for interaction of factor A and t distribution
factor B characteristics, 251– 252
SSB: see Sum o f squares for factor B defined, 251
SSE ; see Error sum o f squares degrees of freedom, 251, 252
SSR: see Regression sum o f squares hypothesis testing, 288– 291
SST; see Total sum of squares t table, 252– 253, 648–649
SSTR; see Sum o f squares due to treatments t test
Standard & Poor’s, 450 assumptions, 610
Standard deviation two-tailed, 458–461, 467
o f binomial random variables, 152– 153 tdf values, 252– 253
Chebyshev’s theorem, 76– 77, 78 Test of independence, 368–372
defined, 70 Test statistics
of discrete random variable, 142– 143 for differences in means, 309– 310
empirical rule, 77– 78 goodness-of-fit, 363– 364
Jarque-Bera, 376 Unbalanced data, 397–398
for Kruskal-Wallis test, 623 Unbiasedness, 242– 243
for mean difference, 317–318 Unconditional probability, 108, 119
for one-way ANOVA, 391 – 393 Unexplained variation, 444
for population correlation coefficient, 426–427 Union, of events, 97, 98, 105– 106
for population mean, with known standard U.S. Bureau of Labor Statistics, 6, 7, 599
deviation, 278 U.S. Census Bureau, 7, 56
for population proportion, 294 University of Pennsylvania Medical Center, 4
for population variance, 340 Unrestricted models, 468
for population variance ratio, 349 Unsystematic patterns, 548
for proportion differences, 325 Unweighted aggregate price indices, 592–593
for sign test, 632 Upper control limit (UCL), 228–229, 230
for test of independence, 371 USA TODAY, 5
for Wald-Wolfowitz runs test, 635
for Wilcoxon rank-sum test, 618–619
for Wilcoxon signed-rank test, 611–612
V
Vanguard, 53, 73, 335, 352, 449, 450, 609, 628
Time series data; see also Forecasting; Returns; Smoothing
Variability measures; see Dispersion measures
techniques
Variables; see also Dummy variables;
defined, 6, 546
Explanatory variables; Qualitative
deflated, 598–600
variables; Quantitative variables; Random
nominal terms, 598
variables
real terms, 598–600
continuous, 9
serial correlation, 480–481
defined, 8
systematic patterns, 548, 562
discrete, 9
unsystematic patterns, 548
response, 430–431, 437–4 38
Total probability rule, 118– 121, 123
Variance; see also Analysis of variance; Population variance;
Total sum o f squares (SST), 392, 411, 444
Sample variance
Toyota Motor Corp., 230
defined, 70
Transformations; see also Logarithms; Polynomial regression
of discrete random variable, 142– 143
models
for frequency distribution, 80– 81
inverse, 189– 191
normal, 187– 189 mean-variance analysis, 74–75
o f portfolio returns, 146– 148
Trend models
shortcut formula, 71,81
exponential, 556–559, 567– 568
Venn, John, 97
linear, 555–556, 567
Venn diagrams, 97
polynomial, 559–560
quadratic, 559–560
seasonal dummy variables, 567– 569 W
Trendlines, 432 Wald-Wolfowitz runs test, 634– 637
Trends Wall Street Journal, 7
extracting, 565– 566 Websites, data sources, 7
systematic patterns, 548 Weighted aggregate price indices, 593– 596
Tukey, John, 39, 395 Weighted mean, 81
Tukey’s honestly significant differences (HSD) method, Wilcox, R. A., 657, 658
397–399,408 Wilcoxon, F., 657, 658
Two-tailed hypothesis tests, 273–274, 285 Wilcoxon rank-sum test
Two-tailed t test, 458– 461, 467 critical values, 618, 658
Two-way ANOVA for independent samples, 614, 616– 619
defined, 402 software, 619
example, 402–403 test statistic, 618– 619
with interaction, 410–4 1 3 uses of, 312, 614– 615
without interaction, 402–4 0 8 Wilcoxon signed-rank test
randomized block designs, 404–406 critical values, 612, 657
sample sizes, 403 for matched-pairs sample, 615– 616
uses of, 402 for population median, 610– 613
Type I errors, 275–276 software, 616
Type II errors, 275–276 test statistic, 611– 612
uses of, 320, 614–615
U Within-treatments estimate of population
UCL; see Upper control limit variance, 390, 391
Unadjusted seasonal index, 564 Within-treatments variability, 390
X z values
defined, 179– 180
x –bar charts, 227– 229
X 2 distribution; see Chi-square distribution finding for given probability, 182– 184
finding probabilities for, 180– 182
inverse transformation, 189– 191
Y
normal transformation, 187– 189
Yields; see Returns
Zillow.com, 7
z
z table, 179– 180 , 646– 647