/*! \page intro Introduction to Bayesian Decision Making Toolbox BDM This is a brief introduction into elements used in the BDM. The toolbox was designed for two principle tasks: Theoretically, the latter is a special case of the former, however we list it separately to highlight its importance in practical applications. Here, we describe basic objects that are required for implementation of the Bayesian parameter estimation. Key objects are:
Bayesian Model: class \c BM
which is an encapsulation of the likelihood function, the prior and methodology of evaluation of the Bayes rule. This methodology may be either exact or approximate.
Posterior density of the parameter: class \c epdf
representing posterior density of the parameter. Methods defined on this class allow any manipulation of the posterior, such as moment evaluation, marginalization and conditioning.
\section bm Class BM The class BM is designed for both on-line and off-line estimation. We make the following assumptions about data: On-line estimation is implemented by method \code void bayes(vec dt)\endcode Off-line estimation is implemented by method \code void bayes_batch(mat D)\endcode As an intermediate product, the bayes rule computes marginal likelihood of the data records \f$ f(D) \f$. Numerical value of this quantity which is important e.g. for model selection can be obtained by calling method \c _ll(). \section epdf Getting results from BM Class \c BM offers several ways how to obtain results: Underscore in the name of method \c _epdf() indicate that the method returns a pointer to the internal posterior density of the model. On the other hand, \c predictor creates a new structure of type \c epdf(). Direct evaluation of predictive pdfs via logpred offers a shortcut for more efficient implementation. \section epdf Probability densities As introduced above, the results of parameter estimation are in the form of probability density function conditioned on numerical values. This type of information is represented by class \c epdf. This class allows such as moment evaluation via methods \c mean() and \c variance(), marginalization via method \c marginal(), and conditioning via method \c condition(). Also, it allows generation of a sample via \c sample() and evaluation of one value of the posterior parameter likelihood via \c evallog(). Multivariate versions of these operations are also available by adding suffix \c _m, i.e. \c sample_m() and \c evallog_m(). These methods providen multiple samples and evaluation of likelihood in multiple points respectively. \section pc Classes for probability calculus When a more demanding task then generation of point estimate of the parameter is required, the power of general probability claculus can be used. The following classes (together with \c epdf introduced above) form the basis of the calculus: The former class is an extension of mpdf that allows conditioning on a symbolic variable. Hence, when numerical results - such as samples - are required, numericla values of the condition must be provided. The names of methods of the \c epdf are used extended by suffix \c cond, i.e. \c samplecond(), \c evallogcond(), where \c cond precedes matrix estension, i.e. \c samplecond_m() and \c evallogcond_m(). The latter class is used to identify how symbolic variables are to be combined together. For example, consider the task of composition of pdfs via the chain rule: \f[ f(a,b,c) = f(a|b,c) f(b) f(c) \f] In our setup, \f$ f(a|b,c) \f$ is represented by an \c mpdf while \f$ f(b) \f$ and \f$ f(c) \f$ by two \c epdfs. We need to distinguish the latter two from each other and to deside in which order they should be added to the mpdf. This distinction is facilitated by the class \c RV which uniquely identify a random varibale. Therefore, each pdf keeps record on which RVs it represents; \c epdf needs to know only one \c RV stored in the attribute \c rv; \c mpdf needs to keep two \c RVs, one for variable on which it is defined (\c rv) and one for variable incondition which is stored in attribute \c rvc. */