# Fisher Linear Discriminant Analysis

Important

Support for Machine Learning Studio (classic) will end on 31 August 2024. We recommend you transition to Azure Machine Learning by that date.

Beginning 1 December 2021, you will not be able to create new Machine Learning Studio (classic) resources. Through 31 August 2024, you can continue to use the existing Machine Learning Studio (classic) resources.

- See information on moving machine learning projects from ML Studio (classic) to Azure Machine Learning.
- Learn more about Azure Machine Learning.

ML Studio (classic) documentation is being retired and may not be updated in the future.

*Identifies the linear combination of feature variables that can best group data into separate classes*

Category: Feature Selection Modules

Note

**Applies to**: Machine Learning Studio (classic) *only*

Similar drag-and-drop modules are available in Azure Machine Learning designer.

## Module overview

This article describes how to use the **Fisher Linear Discriminant Analysis** module in Machine Learning Studio (classic), to create a new feature dataset that captures the combination of features that best separates two or more classes.

This method is often used for dimensionality reduction, because it projects a set of features onto a smaller feature space while preserving the information that discriminates between classes. This not only reduces computational costs for a given classification task, but can help prevent overfitting.

To generate the scores, you provide a label column and set of numerical feature columns as inputs. The algorithm determines the optimal combination of the input columns that linearly separates each group of data while minimizing the distances within each group. The module returns a dataset containing the compact, transformed features, along with a transformation that you can save and apply to another dataset.

### More about linear discriminant analysis

Linear discriminant analysis is similar to analysis of variance (ANOVA) in that it works by comparing the means of the variables. Like ANOVA, it relies on these assumptions:

- Predictors are independent
- The conditional probability density functions of each sample are normally distributed
- Variances among groups are similar

Linear Discriminant Analysis is sometimes abbreviated to LDA, but this is easily confused with *Latent Dirichlet Allocation*. The techniques are completely different, so in this documentation, we use the full names wherever possible.

## How to configure Linear Discriminant Analysis

Add your input dataset and check that the input data meets these requirements:

- Your data should be as complete as possible. Rows with any missing values are ignored.
- Values are expected to have a normal distribution. Before using
**Fisher Linear Discriminant Analysis**, review the data for outliers, or test the distribution. - You should have fewer predictors than there are samples.
- Remove any non-numeric columns. The algorithm examines
**all**valid numeric columns included in the inputs, and return an error if invalid columns are included. If you need to exclude any numeric columns, add a Select Columns in Dataset module before**Fisher Linear Discriminant Analysis**, to create a view that contains only the columns you wish to analyze. You can rejoin the columns later using Add Columns. The original order of rows is preserved.

Connect the input data to the

**Fisher Linear Discriminant Analysis**module.For

**Class labels column**, click**Launch column selector**and choose one label column.For

**Number of feature extractors**, type the number of columns that you want as a result.For example, if your dataset contains eight numeric feature columns, you might type

`3`

to collapse them into a new, reduced feature space of only three columns.It is important to understand that the output columns do not correspond exactly to the input columns, but rather represent a compact transformation of the values in the input columns.

If you use 0 as the value for

**Number of feature extractors**, and*n*columns are used as input,*n*feature extractors are returned, containing new values representing the n-dimensional feature space.Run the experiment.

### Results

The algorithm determines the combination of values in the input columns that linearly separates each group of data while minimizing the distances within each group, and creates two outputs:

**Transformed features**. A dataset containing the specified number of feature extractor columns, named**col1**,**col2**,**col3**, and so forth. The output also includes the class or label variable as well.You can use this compact set of values for training a model.

**Fisher linear discriminant analysis transformation**. A transformation that you can save and then apply to a dataset that has the same schema. This is useful if you are analyzing many datasets of the same type and want to apply the same feature reduction to each. The dataset that you apply it to should have the same schema.

## Examples

For examples of feature selection in machine learning, see the Azure AI Gallery:

- Fisher Linear Discriminant Analysis: Demonstrates how to use this module for dimensionality reduction.

## Technical notes

This section contains implementation details, tips, and answers to frequently asked questions.

### Usage tips

This method works only on continuous variables, not categorical or ordinal variables.

Rows with missing values are ignored when computing the transformation matrix.

If you save a transformation from an experiment, the transformations computed from the original experiment are reapplied to each new set of data, and are not recomputed. Therefore, if you want to compute a new feature set for each set of data, use a new instance of

**Fisher Linear Discriminant Analysis**for each dataset.

### Implementation details

The dataset of features is transformed using *eigenvectors*. The eigenvectors for the input dataset are computed based on the provided feature columns, also called a *discrimination matrix*.

The transformation output by the module contains these eigenvectors, which can be applied to transform another dataset that has the same schema.

For more information about how the eigenvalues are calculated, see this paper (PDF): Eigenvector-based Feature Extraction for Classification. Tymbal, Puuronen et al.

## Expected inputs

Name | Type | Description |
---|---|---|

Dataset | Data Table | Input dataset |

## Module parameters

Name | Type | Range | Optional | Default | Description |
---|---|---|---|---|---|

Class labels column | ColumnSelection | Required | None | Select the column that contains the categorical class labels | |

Number of feature extractors | Integer | >=0 | Required | 0 | Number of feature extractors to use. If zero, then all feature extractors will be used |

## Outputs

Name | Type | Description |
---|---|---|

Transformed features | Data Table | Fisher linear discriminant analysis features transformed to eigenvector space |

Fisher linear discriminant analysis transformation | ITransform interface | Transformation of Fisher linear discriminant analysis |

## Exceptions

Exception | Description |
---|---|

Error 0001 | Exception occurs if one or more specified columns of data set couldn't be found. |

Error 0003 | Exception occurs if one or more of inputs are null or empty. |

Error 0017 | Exception occurs if one or more specified columns have type unsupported by current module. |

For a list of errors specific to Studio (classic) modules, see Machine Learning Error codes.

For a list of API exceptions, see Machine Learning REST API Error Codes.

## See also

Feature Selection

Filter Based Feature Selection

Principal Component Analysis