Bayesian Data Analysis for Animal Scientists

Bayesian Data Analysis for Animal Scientists

The Basics

Blasco, Agustin

Springer International Publishing AG

08/2018

275

Mole

Inglês

9783319853598

15 a 20 dias

615

Descrição não disponível.
Foreword



Notation



1. Do we understand classical statistics?

1.1. Historical introduction

1.2. Test of hypothesis

1.2.1. The procedure

1.2.2. Common misinterpretations

1.3. Standard errors and Confidence intervals

1.3.1. Definition of standard error and confidence interval

1.3.2. Common misinterpretations

1.4. Bias and Risk of an estimator

1.4.1. Unbiased estimators

1.4.2. Common misinterpretations

1.5. Fixed and random effects

1.5.1. Definition of "fixed" and "random" effects

1.5.2. Shrinkage of random effects estimates

1.5.3. Bias, variance and Risk of an estimator when the effect is fixed or random

1.5.4. Common misinterpretations

1.6. Likelihood

1.6.1. Definition of likelihood

1.6.2. The method of maximum likelihood

1.6.3. Common misinterpretations

Appendix 1.1

Appendix 1.2

Appendix 1.3

Appendix 1.4



2. The Bayesian choice

2.1. Bayesian inference

2.1.1. The foundations of Bayesian inference

2.1.2. Bayes theorem

2.1.3. Prior information

2.2. Features of Bayesian inference

2.2.1. Point estimates: Mean, median, mode

2.2.2. Credibility intervals

2.2.3. Marginalisation

2.3. Test of hypotheses

2.3.1. Model choice

2.3.2. Bayes factors

2.3.3. Model averaging

2.4. Common misinterpretations

2.5. Bayesian Inference in practice

2.6. Advantages of Bayesian inference

Appendix 2.1

Appendix 2.2

Appendix 2.3



3. Posterior distributions

3.1. Notation

3.2. Probability density function

3.2.1. Definition

3.2.2. Transformed densities

3.3. Features of a distribution

3.3.1. Mean

3.3.2. Median

3.3.3. Mode

3.3.4. Credibility intervals

3.4. Conditional distribution

3.4.1. Bayes Theorem

3.4.2. Conditional distribution of the sample of a Normal distribution

3.4.3. Conditional distribution of the variance of a Normal distribution

3.4.4. Conditional distribution of the mean of a Normal distribution

3.5. Marginal distribution

3.5.1. Definition

3.5.2. Marginal distribution of the variance of a normal distribution

3.5.3. Marginal distribution of the mean of a normal distribution

Appendix 3.1

Appendix 3.2

Appendix 3.3

Appendix 3.4



4. MCMC

4.1. Samples of Marginal Posterior distributions

4.1.1. Taking samples of Marginal Posterior distributions

4.1.2. Making inferences from samples of Marginal Posterior distributions

4.2. Gibbs sampling

4.2.1. How it works

4.2.2. Why it works

4.2.3. When it works

4.2.4. Gibbs sampling features

4.2.5. Example

4.3. Other MCMC methods

4.3.1. Acceptance-Rejection

4.3.2. Metropolis

Appendix 4.1



5. The "baby" model

5.1. The model

5.2. Analytical solutions

5.2.1. Marginal posterior distribution of the mean and variance

5.2.2. Joint posterior distribution of the mean and variance

5.2.3. Inferences

5.3. Working with MCMC

5.3.1. The process

5.3.2. Using Flat priors

5.3.3. Using vague informative priors

5.3.4. Common misinterpretations

Appendix 5.1

Appendix 5.2

Appendix 5.3



6. The linear model. I. The "fixed" effects model

6.1. The model

6.1.1. The model

6.1.2. Example

6.1.3. Common misinterpretations

6.2. Marginal posterior distributions via MCMC using Flat priors

6.2.1. Joint posterior distribution

6.2.2. Conditional distributions

6.2.3. Gibbs sampling

6.3. Marginal posterior distributions via MCMC using vague informative priors

6.3.1. Vague in formative priors

6.3.2. Conditional distributions.

6.4. Least Squares as a Bayesian Estimator

Appendix 6.1

Appendix 6.2



7. The linear model. II. The "mixed" model

7.1. The mixed model with repeated records

7.1.1. The model

7.1.2. Common misinterpretations

7.1.3. Marginal posterior distributions via MCMC

7.1.4. Gibbs sampling

7.2. The genetic animal model

7.2.1. The model

< 7.2.2. Marginal posterior distributions via MCMC

7.3. Bayesian interpretation of BLUP and REML

7.3.1. BLUP in a frequentist context

7.3.2. BLUP as a Bayesian estimator

7.3.3. REML as a Bayesian estimator

7.4. The multitrait model

7.4.1. The model

7.4.2. Data augmentation

7.4.3. More complex models

Appendix 7.1



8. A scope of the possibilities of Bayesian inference + MCMC

8.1. Nested models: Examples in growth curves

8.1.1. The model

8.1.2. Marginal posterior distributions

8.1.3. More complex models

8.2. Modelling residuals. Examples in canalization.

8.2.1. The model

8.2.2. Marginal posterior distributions

8.2.3. More complex models

8.3. Modelling priors: Examples in genomic selection

8.3.1. The model

8.3.2. RR-BLUP

8.3.3. Bayes A

8.3.4. Bayes B

8.3.5. Bayes C and Bayes Cp

8.3.6. Bayes L (Bayesian Lasso)

8.3.7. Bayesian alphabet in practice

Appendix 8.1



9. Prior information

Abstract

9.1. Exact prior information

9.1.1. Prior information

9.1.2. Posterior probabilities with exact prior information

9.1.3. Influence of prior information in posterior probabilities

9.2. Vague prior information

9.2.1. A vague definition of vague prior information

9.2.2. Examples of the use of vague prior information

9.3. No prior information

9.3.1. Flat priors

9.3.2. Jeffrey's priors

9.3.3. Bernardo's "Reference" priors

9.4. Improper priors

9.5. The Achilles heel of Bayesian inference

Appendix 9.1

Appendix 9.2



10. Model choice

10.1 Model selection

10.1.1. The purpose of model selection

10.1.2. Fitting data vs predicting new records

10.1.3. Common misinterpretations

10.2. Hypothesis tests

10.2.1. Likelihood ratio test and other frequentist tests

10.2.2. Bayesian model choice

10.3. The concept of Information

10.3.1. Fisher information

10.3.2. Shannon information and entropy

10.3.3. Kullback Information and Divergence

10.4. Model selection criteria

10.4.1. Akaike Information Criterion (AIC)

10.4.2. Deviance Information Criterion (DIC)

10.4.3. Bayesian Information Criterion (BIC)

10.4.4. Model choice in practice

Appendix 10.1

Appendix 10.2

Appendix 10.3

Appendix 10.4



Appendix: Three new dialogues between Hylas and Filonus on scientific inference



References

?
Este título pertence ao(s) assunto(s) indicados(s). Para ver outros títulos clique no assunto desejado.
Bayesian statistics;Animal production;Animal breeding;Biostatistics;MCMC, Monte-Carlo Markov Chain methods