Implements algorithms for learning discrete Bayesian network classifiers from data, as well as functions for using these classifiers for prediction, assessing their predictive performance, and inspecting and analyzing their properties.
ExampleLoad a data set and learn a one-dependence estimator by maximizing Akaikeâs information criterion (AIC) score.
After learning the networkâs parameters, you can use it to classify data.
tn <- lp(tn, car, smooth = 0.01)
p <- predict(tn, car, prob = TRUE)
head(p)
#> unacc acc good vgood
#> [1,] 1 3.963694e-09 5.682130e-09 4.269700e-09
#> [2,] 1 1.752769e-09 3.310473e-12 3.236335e-09
#> [3,] 1 3.730170e-09 1.090296e-08 1.800719e-12
#> [4,] 1 3.963694e-09 5.682130e-09 4.269700e-09
#> [5,] 1 4.562294e-09 6.965323e-09 4.536532e-09
#> [6,] 1 4.281155e-09 5.366306e-09 5.168828e-09
p <- predict(tn, car, prob = FALSE)
head(p)
#> [1] unacc unacc unacc unacc unacc unacc
#> Levels: unacc acc good vgood
Estimate predictive accuracy with cross validation.
Or compute the log-likelihood
InstallMake sure you have at least version 3.2.0 of R. You can install bnclassify
from CRAN:
Or get the current development version from Github:
Ideally, you would use the build_vignettes = TRUE
version, and thus get the vignettes, but it requires programs such as texi2dvi to be installed on your side.
See an overview of the package and examples of usage:
See the list of main functionalities.
Use the usage vignette for more details on the functions.
Then have a look at the remaining vignettes.
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4