Adversarial random forests (ARFs) recursively partition data into fully factorized leaves, where features are jointly independent. The procedure is iterative, with alternating rounds of generation and discrimination. Data become increasingly realistic at each round, until original and synthetic samples can no longer be reliably distinguished. This is useful for several unsupervised learning tasks, such as density estimation and data synthesis. Methods for both are implemented in this package. ARFs naturally handle unstructured data with mixed continuous and categorical covariates. They inherit many of the benefits of RFs, including speed, flexibility, and solid performance with default parameters.
InstallationThe arf
package is available on CRAN
:
To install the development version from GitHub using devtools
, run:
devtools::install_github("bips-hb/arf")
Examples
Using Fisherâs iris dataset, we train an ARF and estimate distribution parameters:
# Train the ARF
arf <- adversarial_rf(iris)
# Estimate distribution parameters
psi <- forde(arf, iris)
Density estimation
To estimate log-likelihoods:
mean(lik(arf, psi, iris))
Generative modeling
To generate 100 synthetic samples:
Conditional expectationsTo estimate the mean of some variable(s), optionally conditioned on some event(s):
evi <- data.frame(Species = "setosa")
expct(psi, query = "Sepal.Length", evidence = evi)
For more detailed examples, see the package vignette.
Python libraryA Python implementation of ARF, arfpy
, is available on PyPI. For the development version, see here.
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4