skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: IN-SAMPLE ASYMPTOTICS AND ACROSS-SAMPLE EFFICIENCY GAINS FOR HIGH FREQUENCY DATA STATISTICS
We revisit in-sample asymptotic analysis extensively used in the realized volatility literature. We show that there are gains to be made in estimating current realized volatility from considering realizations in prior periods. The weighting schemes also relate to Kalman-Bucy filters, although our approach is non-Gaussian and model-free. We derive theoretical results for a broad class of processes pertaining to volatility, higher moments, and leverage. The paper also contains a Monte Carlo simulation study showing the benefits of across-sample combinations.  more » « less
Award ID(s):
2015544
PAR ID:
10404661
Author(s) / Creator(s):
; ;
Date Published:
Journal Name:
Econometric Theory
Volume:
39
Issue:
1
ISSN:
0266-4666
Page Range / eLocation ID:
70 to 106
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Andersen, Torben; Chen, Xiaohong (Ed.)
    This paper introduces a unified approach for modeling high-frequency financial data that can accommodate both the continuous-time jump–diffusion and discrete-time realized GARCH model by embedding the discrete realized GARCH structure in the continuous instantaneous volatility process. The key feature of the proposed model is that the corresponding conditional daily integrated volatility adopts an autoregressive structure, where both integrated volatility and jump variation serve as innovations. We name it as the realized GARCH-Itô model. Given the autoregressive structure in the conditional daily integrated volatility, we propose a quasi-likelihood function for parameter estimation and establish its asymptotic properties. To improve the parameter estimation, we propose a joint quasi-likelihood function that is built on the marriage of daily integrated volatility estimated by high-frequency data and nonparametric volatility estimator obtained from option data. We conduct a simulation study to check the finite sample performance of the proposed methodologies and an empirical study with the S&P500 stock index and option data. 
    more » « less
  2. null (Ed.)
    It is important to collect credible training samples $(x,y)$ for building data-intensive learning systems (e.g., a deep learning system). Asking people to report complex distribution $p(x)$, though theoretically viable, is challenging in practice. This is primarily due to the cognitive loads required for human agents to form the report of this highly complicated information. While classical elicitation mechanisms apply to eliciting a complex and generative (and continuous) distribution $p(x)$, we are interested in eliciting samples $$x_i \sim p(x)$$ from agents directly. We coin the above problem sample elicitation. This paper introduces a deep learning aided method to incentivize credible sample contributions from self-interested and rational agents. We show that with an accurate estimation of a certain $$f$$-divergence function we can achieve approximate incentive compatibility in eliciting truthful samples. We then present an efficient estimator with theoretical guarantees via studying the variational forms of the $$f$$-divergence function. We also show a connection between this sample elicitation problem and $$f$$-GAN, and how this connection can help reconstruct an estimator of the distribution based on collected samples. Experiments on synthetic data, MNIST, and CIFAR-10 datasets demonstrate that our mechanism elicits truthful samples. Our implementation is available at https://github.com/weijiaheng/Credible-sample-elicitation.git. 
    more » « less
  3. We present a novel inference approach that we call sample out-of-sample inference. The approach can be used widely, ranging from semisupervised learning to stress testing, and it is fundamental in the application of data-driven distributionally robust optimization. Our method enables measuring the impact of plausible out-of-sample scenarios in a given performance measure of interest, such as a financial loss. The methodology is inspired by empirical likelihood (EL), but we optimize the empirical Wasserstein distance (instead of the empirical likelihood) induced by observations. From a methodological standpoint, our analysis of the asymptotic behavior of the induced Wasserstein-distance profile function shows dramatic qualitative differences relative to EL. For instance, in contrast to EL, which typically yields chi-squared weak convergence limits, our asymptotic distributions are often not chi-squared. Also, the rates of convergence that we obtain have some dependence on the dimension in a nontrivial way but remain controlled as the dimension increases. 
    more » « less
  4. Energy storage can generate significant revenue by taking advantage of fluctuations in real-time energy market prices. In this paper, we investigate the real-time price arbitrage potential of aerodynamic energy storage in wind farms. This under-explored source of energy storage can be realized by deferring energy extraction by turbines toward the front of a farm for later extraction by downstream turbines. In large wind farms, this kinetic energy can be stored for minutes to tens of minutes, depending on the inter-turbine travel distance and the incoming wind speed. This storage mechanism requires minimal capital costs for implementation and potentially could provide additional revenue to wind farm operators. We demonstrate that the potential for revenue generation depends on the energy arbitrage (storage) efficiency and the wind travel time between turbines. We then characterize how price volatility and arbitrage efficiency affect real-time energy market revenue potential. Simulation results show that when price volatility is low, which is the historic norm, noticeably increased revenue is only achieved with high arbitrage efficiencies. However, as price volatility increases, which is expected in the future as the composition of the power system evolves, revenues increase by several percent. 
    more » « less
  5. The assumption that training and testing samples are generated from the same distribution does not always hold for real-world machine-learning applications. The procedure of tackling this discrepancy between the training (source) and testing (target) domains is known as domain adaptation. We propose an unsupervised version of domain adaptation that considers the presence of only unlabelled data in the target domain. Our approach centres on finding correspondences between samples of each domain. The correspondences are obtained by treating the source and target samples as graphs and using a convex criterion to match them. The criteria used are first-order and second-order similarities between the graphs as well as a class-based regularization. We have also developed a computationally efficient routine for the convex optimization, thus allowing the proposed method to be used widely. To verify the effectiveness of the proposed method, computer simulations were conducted on synthetic, image classification and sentiment classification datasets. Results validated that the proposed local sample-to- sample matching method out-performs traditional moment-matching methods and is competitive with respect to current local domain-adaptation methods. 
    more » « less