% LaTeX source for review in J. Sci. Explor.

\documentclass{article}

\begin{document}

\noindent
\textbf{Bayesian Statistics: An Introduction} (3rd ed.) by Peter M.\ Lee. 
London:  Arnold Publishers, 2004. 351  + xv  pp. \$17.95. ISBN
0340814055.  

\bigskip

Bayesian statistics has made great strides in recent years, due partly
to better  understanding of  priors  (e.g., automatic or reference
priors that can  be used in  the absence of subjective prior
information), partly due to the introduction  of  Markov Chain Monte
Carlo (MCMC) techniques for drawing samples  from the  posterior
distribution even when the sample space is huge, and partly due to the 
power  of  hierarchical Bayes models for  describing  very  complex
problems in  a  natural way. Added to this, the result  of  a  Bayesian
analysis naturally provides what scientists really would like to
know, whereas the interpretation  of  the results  of  a  standard
frequentist analysis is often unnatural and  confusing,  especially to
working scientists, but  even  to many  of  those with statistical
training. No longer is the Bayesian approach regarded with suspicion or
disdain  by  most classically trained statisticians; rather, some  of 
the most cutting-edge  work in statistics is being done by statisticians
 of  all stripes using the Bayesian  point-of-view. This is a tribute to
the power of  modern Bayesian methods.  The present book is the third
update of  a book that has become something of  a standard in
introductory texts on the subject.  It is not much different from the 
second edition, the  new material being a chapter on hierarchical Bayes
and an  expansion  of  the chapter  on MCMC (which was added in the
second edition).  However, the basic approach and most of the topics
remain virtually unchanged  from the second edition.  

I personally learned a great deal from the earlier editions of this book
when I  was discovering the power of Bayesian methods in my own work.
The selection  of  topics is basic, including chapters on inference for
normally distributed data  and for data having other distributions 
(e.g., binomial, Poisson and other sorts),  hypothesis testing (an area
in which the numerical results can differ substantially  from those
provided  by  standard hypothesis tests, particularly  for  point null 
hypotheses), two-sample problems, and Bayesian results on correlation,
regression and analysis of variance. A chapter on ``other topics''
discusses some of the  salient  features  of  Bayesian analysis that 
differ from  standard statistical  discussions, such as the important
Likelihood Principle (which standard methods  often violate), the
Stopping Rule Principle (which under mild conditions insulates  Bayesian
methods from  the  problems that arise in standard hypothesis testing 
when  the experimenter  is  allowed to  stop  optionally and
conditionally  on  the  results so far) and the role of decision theory.
The chapter on hierarchical Bayes  discusses the surprising Stein 
``paradox,'' whereby the obvious  estimator  for  a vector parameter under
square-error loss is inadmissible in classical decision  theory; better
estimators can be found naturally using a hierarchical Bayes model. 
Finally, a chapter  on MCMC  rounds out  the  book; MCMC  has 
revolutionized  Bayesian statistics over the past fifteen years by
providing a practical method for  obtaining results, especially
integrals,  by  posterior simulation. (In spirit these
interpretation  is  different: Standard statistics regards parameters as
fixed and  data  as  random variables, and averages or simulates over
the data, whereas  Bayesian statistics regards parameters as random
variables and the observed data  as fixed, and averages or simulates
over the parameters.) The chapter on MCMC  introduces most  of  the
standard techniques (with slice sampling  as a  notable  exception)
using examples coded in the free computer language R.  

The book has much to recommend it, but I do have some problems with it.
For  one thing, much  of  the  book  is devoted to exact or asymptotic
results, often  using so-called conjugate priors that with standard
error distributions produce  posterior distributions that remain in  the
conjugate family and are analytically  tractable. For example, using a
normal prior on the location variable in  normal  inference  problems 
with  a  known  data  variance results in  a  normal posterior 
distribution, so results are easily calculated. But, for all the
importance  of  such analytical techniques, they are too restrictive for
the vast majority of real-world  problems, which generally require
posterior simulation using MCMC to get  practical results. A student
reading this book might get an exaggerated idea of  the role that these
analytical techniques  play  in practical problems and might  regard
MCMC techniques  as  an  afterthought, whereas the truth is quite the 
opposite. It is for this reason that I no longer use it as a textbook in
my Bayesian  course, which  I  teach using simulation  as  the main
calculational technique,  introducing exact analytical results after
discussing problems  from a  simulation point of  view. My aim has
been to prepare the students to tackle real-world  problems in their
chosen field after this one-semester graduate course. The post-course
experience  of  my  students attests to the success  of this approach. 

Thus, I am of  two minds when recommending this book. Certainly I
learned  much from the first  edition,  so  it  can  be  useful  for 
self-study  by  a  mature scientist who is aware of its limitations; but
I would be careful about using it as  a textbook in a course, at least
without balancing it with other material (e.g., the  highly-regarded but
more advanced  book by  Gelman, Carlin, Stem and Rubin)  or with
lectures that placed more emphasis on simulation methods. 

\begin{flushright}
\textit{WILLIAM H.\ JEFFERYS \\
Harlan J.\ Smith Centennial Professor in Astronomy, Emeritus \\
Department of  Astronomy \\
1 University Station, C1400 \\
University  of  Texas at Austin \\ 
Austin, TX 78712-0259}
\end{flushright}

\noindent
\textit{Journal of Scientific Exploration}
\textbf{19} (1) (2005), 131--133.

\end{document}

%