Class 5
February 26th, 2020
typically data collection/integration activities yield high-dimensional data.
Hypergeometry is counter-intuitive and invalidates distance-based analysis
several methods seek to reduce dimensionality: SVD, PCA, LDA
Several methods leverage linear dependencies in data
In bear: chest girth vs. lenght:
In deer: size of fetuses:
n datapoints each having (up to) d dimensions: the data space
A function
Points in the feature space will have new and extra dimensions that will reveal non-linear dependencies. Classifiers will be able to better allocate data points.
However operating with
The new distances in feature space will be accounted for in the Kernel Matrix
The kernel trick
Sometimes all the needed operations in feature space, e.g, distance and average, can be mapped to simple (but n x n) Kernel matrix look-ups!
Thanks to the Kernel trick, we can now forget the original dataset and work with the kernel matrix.
distances get amplified: on third dim. now
Input: a dataset
Compute the
Is K a positive semidefinite matrix?
Apply eigevalue decomposition
How many non-zero e-values? That is the new dimensionality.
define the Mercer mapping
verify that in Mercer space inner products can be mapped to look-ups to the K matrix (see point 2)
where
And so on
Yes,