论文标题

调查:减少数据的几何基础

Survey: Geometric Foundations of Data Reduction

论文作者

Ju, Ce

论文摘要

这项调查是在2016年夏季撰写的。该调查的目的是简要介绍数据降低非线性降低(NLDR)。前两个NLDR分别于2000年发表在科学上,其中解决了具有内在非线性结构的高维数据的类似还原问题。固有的非线性结构始终被解释为计算机科学家和理论物理学家在理论数学中的几何学和拓扑中的概念。在2001年,流形学习的概念首先是一种名为laplacian eigenmaps的NLDR方法。在典型的多种学习设置中,数据集(也称为观察集)分布在RD中嵌入的低维歧管M上,这使每个观察值都具有D维表示。多种学习的目的是将这些观察结果减少为基于几何信息的紧凑型低维度表示。还原过程称为光谱歧管学习。在本文中,我们使用矩阵和操作员表示来得出每个光谱歧管学习,然后我们讨论每种方法中每种方法的收敛行为。因此,该调查被命名为减少数据的几何基础。

This survey is written in summer, 2016. The purpose of this survey is to briefly introduce nonlinear dimensionality reduction (NLDR) in data reduction. The first two NLDR were respectively published in Science in 2000 in which they solve the similar reduction problem of high-dimensional data endowed with the intrinsic nonlinear structure. The intrinsic nonlinear structure is always interpreted as a concept in manifolds from geometry and topology in theoretical mathematics by computer scientists and theoretical physicists. In 2001, the concept of Manifold Learning first appears as an NLDR method called Laplacian Eigenmaps. In a typical manifold learning setup, the data set, also called the observation set, is distributed on or near a low dimensional manifold M embedded in RD, which yields that each observation has a D-dimensional representation. The goal of manifold learning is to reduce these observations as a compact lower-dimensional representation based on the geometric information. The reduction procedure is called the spectral manifold learning. In this paper, we derive each spectral manifold learning with the matrix and operator representation, and we then discuss the convergence behavior of each method in a geometric uniform language. Hence, the survey is named Geometric Foundations of Data Reduction.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源