skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Multiscale regression on unknown manifolds
We consider the regression problem of estimating functions on $$ \mathbb{R}^D $$ but supported on a $ d $-dimensional manifold $$ \mathcal{M} ~~\subset \mathbb{R}^D $$ with $$ d \ll D $$. Drawing ideas from multi-resolution analysis and nonlinear approximation, we construct low-dimensional coordinates on $$ \mathcal{M} $$ at multiple scales, and perform multiscale regression by local polynomial fitting. We propose a data-driven wavelet thresholding scheme that automatically adapts to the unknown regularity of the function, allowing for efficient estimation of functions exhibiting nonuniform regularity at different locations and scales. We analyze the generalization error of our method by proving finite sample bounds in high probability on rich classes of priors. Our estimator attains optimal learning rates (up to logarithmic factors) as if the function was defined on a known Euclidean domain of dimension $ d $, instead of an unknown manifold embedded in $$ \mathbb{R}^D $$. The implemented algorithm has quasilinear complexity in the sample size, with constants linear in $ D $ and exponential in $ d $. Our work therefore establishes a new framework for regression on low-dimensional sets embedded in high dimensions, with fast implementation and strong theoretical guarantees.  more » « less
Award ID(s):
1818751 2012652 2145167
PAR ID:
10296149
Author(s) / Creator(s):
; ;
Date Published:
Journal Name:
Mathematics in Engineering
Volume:
4
Issue:
4
ISSN:
2640-3501
Page Range / eLocation ID:
1 to 25
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract The skew mean curvature flow is an evolution equation for a $$d$$ dimensional manifold immersed into $$\mathbb {R}^{d+2}$$, and which moves along the binormal direction with a speed proportional to its mean curvature. In this article, we prove small data global regularity in low-regularity Sobolev spaces for the skew mean curvature flow in dimensions $$d\geq 4$$. This extends the local well-posedness result in [7]. 
    more » « less
  2. Abstract The skew mean curvature flow is an evolution equation forddimensional manifolds embedded in$${{\mathbb {R}}}^{d+2}$$ R d + 2 (or more generally, in a Riemannian manifold). It can be viewed as a Schrödinger analogue of the mean curvature flow, or alternatively as a quasilinear version of the Schrödinger Map equation. In this article, we prove small data local well-posedness in low-regularity Sobolev spaces for the skew mean curvature flow in dimension$$d\ge 4$$ d 4
    more » « less
  3. Abstract Let M be a complete Riemannian manifold and suppose {p\in M} . For each unit vector {v\in T_{p}M} , the Jacobi operator , {\mathcal{J}_{v}:v^{\perp}\rightarrow v^{\perp}} is the symmetric endomorphism, {\mathcal{J}_{v}(w)=R(w,v)v} . Then p is an isotropic point if there exists a constant {\kappa_{p}\in{\mathbb{R}}} such that {\mathcal{J}_{v}=\kappa_{p}\operatorname{Id}_{v^{\perp}}} for each unit vector {v\in T_{p}M} . If all points are isotropic, then M is said to be isotropic; it is a classical result of Schur that isotropic manifolds of dimension at least 3 have constant sectional curvatures. In this paper we consider almost isotropic manifolds , i.e. manifolds having the property that for each {p\in M} , there exists a constant {\kappa_{p}\in\mathbb{R}} such that the Jacobi operators {\mathcal{J}_{v}} satisfy {\operatorname{rank}({\mathcal{J}_{v}-\kappa_{p}\operatorname{Id}_{v^{\perp}}}% )\leq 1} for each unit vector {v\in T_{p}M} . Our main theorem classifies the almost isotropic simply connected Kähler manifolds, proving that those of dimension {d=2n\geqslant 4} are either isometric to complex projective space or complex hyperbolic space or are totally geodesically foliated by leaves isometric to {{\mathbb{C}}^{n-1}} . 
    more » « less
  4. We study the $$\ell_p$$ regression problem, which requires finding $$\mathbf{x}\in\mathbb R^{d}$$ that minimizes $$\|\mathbf{A}\mathbf{x}-\mathbf{b}\|_p$$ for a matrix $$\mathbf{A}\in\mathbb R^{n \times d}$$ and response vector $$\mathbf{b}\in\mathbb R^{n}$$. There has been recent interest in developing subsampling methods for this problem that can outperform standard techniques when $$n$$ is very large. However, all known subsampling approaches have run time that depends exponentially on $$p$$, typically, $$d^{\mathcal{O}(p)}$$, which can be prohibitively expensive. We improve on this work by showing that for a large class of common \emph{structured matrices}, such as combinations of low-rank matrices, sparse matrices, and Vandermonde matrices, there are subsampling based methods for $$\ell_p$$ regression that depend polynomially on $$p$$. For example, we give an algorithm for $$\ell_p$$ regression on Vandermonde matrices that runs in time $$\mathcal{O}(n\log^3 n+(dp^2)^{0.5+\omega}\cdot\text{polylog}\,n)$$, where $$\omega$$ is the exponent of matrix multiplication. The polynomial dependence on $$p$$ crucially allows our algorithms to extend naturally to efficient algorithms for $$\ell_\infty$$ regression, via approximation of $$\ell_\infty$$ by $$\ell_{\mathcal{O}(\log n)}$$. Of practical interest, we also develop a new subsampling algorithm for $$\ell_p$$ regression for arbitrary matrices, which is simpler than previous approaches for $$p \ge 4$$. 
    more » « less
  5. We consider the problem of efficiently approximating and encoding high-dimensional data sampled from a probability distribution ρ in ℝD, that is nearly supported on a d-dimensional set  - for example supported on a d-dimensional manifold. Geometric Multi-Resolution Analysis (GMRA) provides a robust and computationally efficient procedure to construct low-dimensional geometric approximations of  at varying resolutions. We introduce GMRA approximations that adapt to the unknown regularity of , by introducing a thresholding algorithm on the geometric wavelet coefficients. We show that these data-driven, empirical geometric approximations perform well, when the threshold is chosen as a suitable universal function of the number of samples n, on a large class of measures ρ, that are allowed to exhibit different regularity at different scales and locations, thereby efficiently encoding data from more complex measures than those supported on manifolds. These GMRA approximations are associated to a dictionary, together with a fast transform mapping data to d-dimensional coefficients, and an inverse of such a map, all of which are data-driven. The algorithms for both the dictionary construction and the transforms have complexity CDnlogn with the constant C exponential in d. Our work therefore establishes Adaptive GMRA as a fast dictionary learning algorithm, with approximation guarantees, for intrinsically low-dimensional data. We include several numerical experiments on both synthetic and real data, confirming our theoretical results and demonstrating the effectiveness of Adaptive GMRA. 
    more » « less