The Random Subspace Method for Constructing Decision Forests

Looking for more research papers to read, I scanned my Hands-On Machine Learning notes for the many papers that were referenced there. This is one of those papers. These papers are mainly on machine learning and deep learning topics.

Reference The Random Subspace Method for Constructing Decision Forests Paper

Date Created:
Last Edited:
1 35

Introduction

Much of the previous attention on decision trees focuses on the splitting criteria and optimization of tree sizes. The dilemma between over fitting and achieving maximum accuracy is seldom resolved. In this paper, a method to construct a decision tree based classifier is proposed that maintains highest accuracy on training data and improves on generalization accuracy as it grows in complexity. The classifier consists of multiple trees constructed systematically by pseudo-randomly selecting subsets of components of the feature vector, that is, trees constructed in randomly chosen subspaces. The subspace method is compared to single-tree classifiers and other forest construction methods by experimenting on public available datasets, where the method’s superiority is demonstrated.

A persistent problem in using decision tree for classification is how to avoid overfitting a set of training data while achieving maximum accuracy. This author has proposed combining multiple trees constructed in randomly selected subspaces can achieve nearly monotonic increase in generalization accuracy while preserving perfect accuracy on training data.

Notes

Methods for Constructing a Decision Tree

The majority of tree construction methods use linear splits at each internal node. A typical method selects a hyperplane or multuple hyperplanes at each internal node, and samples are assigned to branches representing different regions of the feature space bounded by such hyperplanes. The methods differ by the number of and orientations of hyperplanes. These method can be categoized by the types of splits they produce:

  1. axis-parallel linear splits A threshold is chosen on the values of a particular feature dimension, and samples are assigned to the branches according to whether the corresponding feature values exceed the threshold.
  2. oblique linear splits Samples are addigned to the branches according to which side of a hyperplane or which region bounded by multiple hyperplanes they fall in, but the hyperplanes are not necessarily parallel to any axis of the feature space
  3. piecewise linear splits Branches represent a Voroni tessellation of the feature space. Samples are assigned on nearest-neighbor matching to chosen anchor points.

Within each category, the splitting function can be obtained in many ways. Stopping criteria is a big issue in tree construction. Each splitting function defines a model for projecting classification form the training sampels to unclassified points in the space. From this point of view, it is not surprising to see that, no method could be universally best for an arbitrary, finite training set.

Systematic Construction of a Forest

The author’s method is another example of inproving accuracy through combining the power of multiple classifiers. Here, each classifier is a decision tree, and the combined classifier is called a decision forest. The method relies on an autonomous, pseudorandom procedure to select a small number of dimensions from a given feature space. In each pass, such a selection is made and a subspace is fixed where all points have a constant value (say zero( in the unselected dimensions. A;; samples are projected to this subspace, and a decsion tree is constructed using the projected training samples.

For a forest to achieve better accuracy than individual trees, it is critical there shoould be sufficient independence or dissimilarity between the trees. A simple measure of similarity between two decsion trees can be the amount that their decision regions overlap on.

You can read more about how comments are sorted in this blog post.

User Comments

Insert Math Markup

ESC
About Inserting Math Content
Display Style:

Embed News Content

ESC
About Embedding News Content

Embed Youtube Video

ESC
Embedding Youtube Videos

Embed TikTok Video

ESC
Embedding TikTok Videos

Embed X Post

ESC
Embedding X Posts

Embed Instagram Post

ESC
Embedding Instagram Posts

Insert Details Element

ESC

Example Output:

Summary Title
You will be able to insert content here after confirming the title of the <details> element.

Insert Table

ESC
Customization
Align:
Preview:

Insert Horizontal Rule

#000000

Preview:


View Content At Different Sizes

ESC

Edit Style of Block Nodes

ESC

Edit the background color, default text color, margin, padding, and border of block nodes. Editable block nodes include paragraphs, headers, and lists.

#ffffff
#000000

Edit Selected Cells

Change the background color, vertical align, and borders of the cells in the current selection.

#ffffff
Vertical Align:
Border
#000000
Border Style:

Edit Table

ESC
Customization:
Align:

Upload Files

ESC

Upload a .lexical file. If the file type matches the type of the current editor, then a preview will be shown below the file input.

Upload Jupyter Notebook

ESC

Upload a Jupyter notebook and embed the resulting HTML in the text editor.

Insert Custom HTML

ESC

Edit Image

ESC
#ffffff

Insert Columns Layout

ESC
Column Type:

Select Code Language

ESC
Select Coding Language

Upload Previous Version of Editor State

ESC