skip to main content

Search for: All records

Creators/Authors contains: "Yue, Y."

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. In this paper we consider the training stability of recurrent neural networks (RNNs) and propose a family of RNNs, namely SBO-RNN, that can be formulated using stochastic bilevel optimization (SBO). With the help of stochastic gradient descent (SGD), we manage to convert the SBO problem into an RNN where the feedforward and backpropagation solve the lower and upper-level optimization for learning hidden states and their hyperparameters, respectively. We prove that under mild conditions there is no vanishing or exploding gradient in training SBO-RNN. Empirically we demonstrate our approach with superior performance on several benchmark datasets, with fewer parameters, less training data, and much faster convergence. Code is available at
  2. Abstract Magnetic fields have an important role in the evolution of interstellar medium and star formation 1,2 . As the only direct probe of interstellar field strength, credible Zeeman measurements remain sparse owing to the lack of suitable Zeeman probes, particularly for cold, molecular gas 3 . Here we report the detection of a magnetic field of +3.8 ± 0.3 microgauss through the H  I narrow self-absorption (HINSA) 4,5 towards L1544 6,7 —a well-studied prototypical prestellar core in an early transition between starless and protostellar phases 8–10 characterized by a high central number density 11 and a low central temperature 12 . A combined analysis of the Zeeman measurements of quasar H  I absorption, H  I emission, OH emission and HINSA reveals a coherent magnetic field from the atomic cold neutral medium (CNM) to the molecular envelope. The molecular envelope traced by the HINSA is found to be magnetically supercritical, with a field strength comparable to that of the surrounding diffuse, magnetically subcritical CNM despite a large increase in density. The reduction of the magnetic flux relative to the mass, which is necessary for star formation, thus seems to have already happened during the transition from the diffuse CNM to the molecular gasmore »traced by the HINSA. This is earlier than envisioned in the classical picture where magnetically supercritical cores capable of collapsing into stars form out of magnetically subcritical envelopes 13,14 .« less
  3. Many sequential decision making tasks can be viewed as combinatorial optimiza- tion problems over a large number of actions. When the cost of evaluating an ac- tion is high, even a greedy algorithm, which iteratively picks the best action given the history, is prohibitive to run. In this paper, we aim to learn a greedy heuris- tic for sequentially selecting actions as a surrogate for invoking the expensive oracle when evaluating an action. In particular, we focus on a class of combinato- rial problems that can be solved via submodular maximization (either directly on the objective function or via submodular surrogates). We introduce a data-driven optimization framework based on the submodular-norm loss, a novel loss func- tion that encourages the resulting objective to exhibit diminishing returns. Our framework outputs a surrogate objective that is efficient to train, approximately submodular, and can be made permutation-invariant. The latter two properties al- low us to prove strong approximation guarantees for the learned greedy heuristic. Furthermore, our model is easily integrated with modern deep imitation learning pipelines for sequential prediction tasks. We demonstrate the performance of our algorithm on a variety of batched and sequential optimization tasks, including set cover, active learning, and data-drivenmore »protein engineering.« less
  4. Adams, RP ; Gogate V (Ed.)
    We study the problem of learning sequential decision-making policies in settings with multiple state-action representations. Such settings naturally arise in many domains, such as planning (e.g., multiple integer programming formulations) and various combinatorial optimization problems (e.g., those with both integer programming and graph-based formulations). Inspired by the classical co-training framework for classification, we study the problem of co-training for policy learning. We present sufficient conditions under which learning from two views can improve upon learning from a single view alone. Motivated by these theoretical insights, we present a meta-algorithm for co-training for sequential decision making. Our framework is compatible with both reinforcement learning and imitation learning. We validate the effectiveness of our approach across a wide range of tasks, including discrete/continuous control and combinatorial optimization.