Skip to contents

Introduction

Adversarial random forests (ARFs) recursively partition data into fully factorized leaves, where features are jointly independent. The procedure is iterative, with alternating rounds of generation and discrimination. Data become increasingly realistic at each round, until original and synthetic samples can no longer be reliably distinguished. This is useful for several unsupervised learning tasks, such as density estimation and data synthesis. Methods for both are implemented in this package. ARFs naturally handle unstructured data with mixed continuous and categorical covariates. They inherit many of the benefits of RFs, including speed, flexibility, and solid performance with default parameters.

Installation

The arf package is available on CRAN:

To install the development version from GitHub using devtools, run:

devtools::install_github("bips-hb/arf")

Examples

Using Fisher’s iris dataset, we train an ARF and estimate distribution parameters:

# Train the ARF
arf <- adversarial_rf(iris)

# Estimate distribution parameters
psi <- forde(arf, iris)

Density estimation

To estimate log-likelihoods:

mean(lik(arf, psi, iris))

Generative modeling

To generate 100 synthetic samples:

forge(psi, 100)

Conditional expectations

To estimate the mean of some variable(s), optionally conditioned on some event(s):

evi <- data.frame(Species = "setosa")
expct(psi, query = "Sepal.Length", evidence = evi)

For more detailed examples, see the package vignette.

Python library

A Python implementation of ARF, arfpy, is available on PyPI. For the development version, see here.

References

  • Watson, D. S., Blesch, K., Kapar, J. & Wright, M. N. (2023). Adversarial random forests for density estimation and generative modeling. In Proceedings of the 26th International Conference on Artificial Intelligence and Statistics. Link here.