Numerical solutions of stochastic problems involving random processes π(π‘), which constitutes infinite families of random variables, require to represent these processes by finite dimensional (FD) models ππ (π‘), i.e., deterministic functions of time depending on finite numbers π of random variables. Most available FD models match the mean, correlation, and other global properties of π(π‘). They provide useful information to a broad range of problems, but cannot be used to estimate extremes or other sample properties of π(π‘). We develop FD models ππ (π‘) for processes π(π‘) with continuous samples and establish conditions under which these models converge weakly to π(π‘) in the space of continuous functions as π β β. These theoretical results are illustrated by numerical examples which show that, under the conditions established in this study, samples and extremes of π(π‘) can be approximated by samples and extremes of ππ (π‘) and that the discrepancy between samples and extremes of these processes decreases with π.
more »
« less
Randomized Dimensionality Reduction for Facility Location and Single-Linkage Clustering
Random dimensionality reduction is a versatile tool for speeding up algorithms for high-dimensional problems. We study its application to two clustering problems: the facility location problem, and the single-linkage hierarchical clustering problem, which is equivalent to computing the minimum spanning tree. We show that if we project the input pointset π onto a random π=π(ππ)-dimensional subspace (where ππ is the doubling dimension of π), then the optimum facility location cost in the projected space approximates the original cost up to a constant factor. We show an analogous statement for minimum spanning tree, but with the dimension π having an extra loglogπ term and the approximation factor being arbitrarily close to 1. Furthermore, we extend these results to approximating solutions instead of just their costs. Lastly, we provide experimental results to validate the quality of solutions and the speedup due to the dimensionality reduction. Unlike several previous papers studying this approach in the context of π-means and π-medians, our dimension bound does not depend on the number of clusters but only on the intrinsic dimensionality of π.
more »
« less
- PAR ID:
- 10276628
- Date Published:
- Journal Name:
- International Conference on Machine Learning
- Page Range / eLocation ID:
- 7948-7957
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Conditions under which samples of continuous stochastic processes π(π‘) on bounded time intervals [0, π] can be represented by samples of finite dimensional (FD) processes ππ (π‘) are augmented such that samples of Slepian models ππ,π(π‘) of ππ (π‘) can be used as surrogates for samples of Slepian models ππ(π‘) of π(π‘). FD processes are deterministic functions of time and π < β random variables. The discrepancy between target and FD samples is quantified by the metric of the space πΆ[0, π] of continuous functions. The numerical illustrations, which include Gaussian/non-Gaussian FD processes and solutions of linear/nonlinear random vibration problems, are consistent with the theoretical findings in the paper.more » « less
-
Consider an instance of Euclidean k-means or k-medians clustering. We show that the cost of the optimal solution is preserved up to a factor of (1+Ξ΅) under a projection onto a random O(log(k /Ξ΅) / Ξ΅2)-dimensional subspace. Further, the cost of every clustering is preserved within (1+Ξ΅). More generally, our result applies to any dimension reduction map satisfying a mild sub-Gaussian-tail condition. Our bound on the dimension is nearly optimal. Additionally, our result applies to Euclidean k-clustering with the distances raised to the p-th power for any constant p. For k-means, our result resolves an open problem posed by Cohen, Elder, Musco, Musco, and Persu (STOC 2015); for k-medians, it answers a question raised by Kannan.more » « less
-
This paper studies differentially private stochastic convex optimization (DP-SCO) in the presence of heavy-tailed gradients, where only a π kth-moment bound on sample Lipschitz constants is assumed, instead of a uniform bound. The authors propose a reduction-based approach that achieves the first near-optimal error rates (up to logarithmic factors) in this setting. Specifically, under ( π , πΏ ) (Ο΅,Ξ΄)-approximate differential privacy, they achieve an error bound of πΊ 2 π + πΊ π β ( π π π ) 1 β 1 π , n β G 2 β β +G k β β ( nΟ΅ d β β ) 1β k 1 β , up to a mild polylogarithmic factor in 1 πΏ Ξ΄ 1 β , where πΊ 2 G 2 β and πΊ π G k β are the 2nd and π kth moment bounds on sample Lipschitz constants. This nearly matches the lower bound established by Lowy and Razaviyayn (2023). Beyond the basic result, the authors introduce a suite of private algorithms that further improve performance under additional assumptions: an optimal algorithm under a known-Lipschitz constant, a near-linear time algorithm for smooth functions, and an optimal linear-time algorithm for smooth generalized linear models.more » « less
-
Abstract Optimal transport (OT) methods seek a transformation map (or plan) between two probability measures, such that the transformation has the minimum transportation cost. Such a minimum transport cost, with a certain power transform, is called the Wasserstein distance. Recently, OT methods have drawn great attention in statistics, machine learning, and computer science, especially in deep generative neural networks. Despite its broad applications, the estimation of highβdimensional Wasserstein distances is a wellβknown challenging problem owing to the curseβofβdimensionality. There are some cuttingβedge projectionβbased techniques that tackle highβdimensional OT problems. Three major approaches of such techniques are introduced, respectively, the slicing approach, the iterative projection approach, and the projection robust OT approach. Open challenges are discussed at the end of the review. This article is categorized under:Statistical and Graphical Methods of Data Analysis > Dimension ReductionStatistical Learning and Exploratory Methods of the Data Sciences > Manifold Learningmore » « less
An official website of the United States government

