Huge product rangeOver 140,000 books & equipment products
Rapid shippingUK & Worldwide
Pay in £, € or U.S.$By card, cheque, transfer, draft
Exceptional customer serviceGet specialist help and advice
Bayesian Data Analysis for Animal Scientists provides an easy introduction to Bayesian inference using Markov chain Monte Carlo (MCMC) techniques, making most topics intuitively reasonable and deriving to appendixes the more complicated matters. The biologist or the agricultural researcher does not normally have a background in Bayesian statistics, having difficulties in following the technical books introducing Bayesian techniques. The difficulties arise from the way of making inferences, which is completely different in the Bayesian school, and from the difficulties in understanding complicated matters such as the MCMC numerical methods. Both schools, classic and Bayesian, are compared here, highlighting the advantages of Bayesian solutions, and proposing inferences based in relevant differences, guaranteed values, probabilities of similitude or the use of ratios. The author also gives a scope of complex problems that can be solved using Bayesian statistics, and ends Bayesian Data Analysis for Animal Scientists by explaining the difficulties associated with model choice and the use of small samples. The book has a practical orientation and uses simple models to introduce the reader in this increasingly popular school of inference.
Foreword Notation 1. Do we understand classical statistics? 1.1. Historical introduction 1.2. Test of hypothesis 1.2.1. The procedure 1.2.2. Common misinterpretations 1.3. Standard errors and Confidence intervals 1.3.1. Definition of standard error and confidence interval 1.3.2. Common misinterpretations 1.4. Bias and Risk of an estimator 1.4.1. Unbiased estimators 1.4.2. Common misinterpretations 1.5. Fixed and random effects 1.5.1. Definition of "fixed" and "random" effects 1.5.2. Shrinkage of random effects estimates 1.5.3. Bias, variance and Risk of an estimator when the effect is fixed or random 1.5.4. Common misinterpretations 1.6. Likelihood 1.6.1. Definition of likelihood 1.6.2. The method of maximum likelihood 1.6.3. Common misinterpretations Appendix 1.1 Appendix 1.2 Appendix 1.3 Appendix 1.4 2. The Bayesian choice 2.1. Bayesian inference 2.1.1. The foundations of Bayesian inference 2.1.2. Bayes theorem 2.1.3. Prior information 2.2. Features of Bayesian inference 2.2.1. Point estimates: Mean, median, mode 2.2.2. Credibility intervals 2.2.3. Marginalisation 2.3. Test of hypotheses 2.3.1. Model choice 2.3.2. Bayes factors 2.3.3. Model averaging 2.4. Common misinterpretations 2.5. Bayesian Inference in practice 2.6. Advantages of Bayesian inference Appendix 2.1 Appendix 2.2 Appendix 2.3 3. Posterior distributions 3.1. Notation 3.2. Probability density function 3.2.1. Definition 3.2.2. Transformed densities 3.3. Features of a distribution 3.3.1. Mean 3.3.2. Median 3.3.3. Mode 3.3.4. Credibility intervals 3.4. Conditional distribution 3.4.1. Bayes Theorem 3.4.2. Conditional distribution of the sample of a Normal distribution 3.4.3. Conditional distribution of the variance of a Normal distribution 3.4.4. Conditional distribution of the mean of a Normal distribution 3.5. Marginal distribution 3.5.1. Definition 3.5.2. Marginal distribution of the variance of a normal distribution 3.5.3. Marginal distribution of the mean of a normal distribution Appendix 3.1 Appendix 3.2 Appendix 3.3 Appendix 3.4 4. MCMC 4.1. Samples of Marginal Posterior distributions 4.1.1. Taking samples of Marginal Posterior distributions 4.1.2. Making inferences from samples of Marginal Posterior distributions 4.2. Gibbs sampling 4.2.1. How it works 4.2.2. Why it works 4.2.3. When it works 4.2.4. Gibbs sampling features 4.2.5. Example 4.3. Other MCMC methods 4.3.1. Acceptance-Rejection 4.3.2. Metropolis Appendix 4.1 5. The "baby" model 5.1. The model 5.2. Analytical solutions 5.2.1. Marginal posterior distribution of the mean and variance 5.2.2. Joint posterior distribution of the mean and variance 5.2.3. Inferences 5.3. Working with MCMC 5.3.1. The process 5.3.2. Using Flat priors 5.3.3. Using vague informative priors 5.3.4. Common misinterpretations Appendix 5.1 Appendix 5.2 Appendix 5.3 6. The linear model. I. The "fixed" effects model 6.1. The model 6.1.1. The model 6.1.2. Example 6.1.3. Common misinterpretations 6.2. Marginal posterior distributions via MCMC using Flat priors 6.2.1. Joint posterior distribution 6.2.2. Conditional distributions 6.2.3. Gibbs sampling 6.3. Marginal posterior distributions via MCMC using vague informative priors 6.3.1. Vague in formative priors 6.3.2. Conditional distributions. 6.4. Least Squares as a Bayesian Estimator Appendix 6.1 Appendix 6.2 7. The linear model. II. The "mixed" model 7.1. The mixed model with repeated records 7.1.1. The model 7.1.2. Common misinterpretations 7.1.3. Marginal posterior distributions via MCMC 7.1.4. Gibbs sampling 7.2. The genetic animal model 7.2.1. The model < 7.2.2. Marginal posterior distributions via MCMC 7.3. Bayesian interpretation of BLUP and REML 7.3.1. BLUP in a frequentist context 7.3.2. BLUP as a Bayesian estimator 7.3.3. REML as a Bayesian estimator 7.4. The multitrait model 7.4.1. The model 7.4.2. Data augmentation 7.4.3. More complex models Appendix 7.1 8. A scope of the possibilities of Bayesian inference + MCMC 8.1. Nested models: Examples in growth curves 8.1.1. The model 8.1.2. Marginal posterior distributions 8.1.3. More complex models 8.2. Modelling residuals. Examples in canalization. 8.2.1. The model 8.2.2. Marginal posterior distributions 8.2.3. More complex models 8.3. Modelling priors: Examples in genomic selection 8.3.1. The model 8.3.2. RR-BLUP 8.3.3. Bayes A 8.3.4. Bayes B 8.3.5. Bayes C and Bayes Cp 8.3.6. Bayes L (Bayesian Lasso) 8.3.7. Bayesian alphabet in practice Appendix 8.1 9. Prior information Abstract 9.1. Exact prior information 9.1.1. Prior information 9.1.2. Posterior probabilities with exact prior information 9.1.3. Influence of prior information in posterior probabilities 9.2. Vague prior information 9.2.1. A vague definition of vague prior information 9.2.2. Examples of the use of vague prior information 9.3. No prior information 9.3.1. Flat priors 9.3.2. Jeffrey's priors 9.3.3. Bernardo's "Reference" priors 9.4. Improper priors 9.5. The Achilles heel of Bayesian inference Appendix 9.1 Appendix 9.2 10. Model choice 10.1 Model selection 10.1.1. The purpose of model selection 10.1.2. Fitting data vs predicting new records 10.1.3. Common misinterpretations 10.2. Hypothesis tests 10.2.1. Likelihood ratio test and other frequentist tests 10.2.2. Bayesian model choice 10.3. The concept of Information 10.3.1. Fisher information 10.3.2. Shannon information and entropy 10.3.3. Kullback Information and Divergence 10.4. Model selection criteria 10.4.1. Akaike Information Criterion (AIC) 10.4.2. Deviance Information Criterion (DIC) 10.4.3. Bayesian Information Criterion (BIC) 10.4.4. Model choice in practice Appendix 10.1 Appendix 10.2 Appendix 10.3 Appendix 10.4 Appendix: Three new dialogues between Hylas and Filonus on scientific inference References â
Agustín Blasco is Professor of Animal Breeding and Genetics Visiting scientist at ABRO (Edinburgh), INRA (Jouy en Josas) and FAO (Rome). He was President of the World Rabbit Science Association and editor in chief of the journal World Rabbit Science. His career has focused on the genetics of litter size components and genetics of meat quality in rabbits and pigs. He has published more than one hundred papers in international journals. Invited speaker several times at the European Association for Animal Production and at the World Congress on Genetics Applied to Livestock Production among others. Chapman Lecturer at the University of Wisconsin. He has taught courses on Bayesian Inference at the universities of Valencia (Spain), Edinburgh (UK), Wisconsin (USA), Padua (Italy), Sao Paulo, Lavras (Brazil), Nacional (Uruguay), Lomas (Argentina) and at INRA in Toulouse (France).