- Home
- Search Results
- Page 1 of 1
Search for: All records
-
Total Resources5
- Resource Type
-
0001000004000000
- More
- Availability
-
41
- Author / Contributor
- Filter by Author / Creator
-
-
Huang, Yu-Jui (5)
-
Wang, Zhenhua (2)
-
Zhou, Zhou (2)
-
Bayraktar, Erhan (1)
-
Guasoni, Paolo (1)
-
Huang, Yu-Chih (1)
-
Lin, Shih-Chun (1)
-
Lin, Wan-Yi (1)
-
Lyu, Kuan-Hui (1)
-
Shen, Hsin-Hua (1)
-
Zhang, Yuchong (1)
-
#Tyler Phillips, Kenneth E. (0)
-
#Willis, Ciara (0)
-
& Abreu-Ramos, E. D. (0)
-
& Abramson, C. I. (0)
-
& Abreu-Ramos, E. D. (0)
-
& Adams, S.G. (0)
-
& Ahmed, K. (0)
-
& Ahmed, Khadija. (0)
-
& Aina, D.K. Jr. (0)
-
- Filter by Editor
-
-
& Spizer, S. M. (0)
-
& . Spizer, S. (0)
-
& Ahn, J. (0)
-
& Bateiha, S. (0)
-
& Bosch, N. (0)
-
& Brennan K. (0)
-
& Brennan, K. (0)
-
& Chen, B. (0)
-
& Chen, Bodong (0)
-
& Drown, S. (0)
-
& Ferretti, F. (0)
-
& Higgins, A. (0)
-
& J. Peters (0)
-
& Kali, Y. (0)
-
& Ruiz-Arias, P.M. (0)
-
& S. Spitzer (0)
-
& Sahin. I. (0)
-
& Spitzer, S. (0)
-
& Spitzer, S.M. (0)
-
(submitted - in Review for IEEE ICASSP-2024) (0)
-
-
Have feedback or suggestions for a way to improve these results?
!
Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
Free, publicly-accessible full text available April 30, 2026
-
Bayraktar, Erhan; Huang, Yu-Jui; Wang, Zhenhua; Zhou, Zhou (, Mathematics of Operations Research)This paper considers an infinite-horizon Markov decision process (MDP) that allows for general nonexponential discount functions in both discrete and continuous time. Because of the inherent time inconsistency, we look for a randomized equilibrium policy (i.e., relaxed equilibrium) in an intrapersonal game between an agent’s current and future selves. When we modify the MDP by entropy regularization, a relaxed equilibrium is shown to exist by a nontrivial entropy estimate. As the degree of regularization diminishes, the entropy-regularized MDPs approximate the original MDP, which gives the general existence of a relaxed equilibrium in the limit by weak convergence arguments. As opposed to prior studies that consider only deterministic policies, our existence of an equilibrium does not require any convexity (or concavity) of the controlled transition probabilities and reward function. Interestingly, this benefit of considering randomized policies is unique to the time-inconsistent case.more » « less
-
Huang, Yu-Jui; Lin, Shih-Chun; Huang, Yu-Chih; Lyu, Kuan-Hui; Shen, Hsin-Hua; Lin, Wan-Yi (, 2023 IEEE International Symposium on Information Theory (ISIT))
-
Huang, Yu-Jui; Zhang, Yuchong (, Journal of machine learning research)
-
Guasoni, Paolo; Huang, Yu-Jui (, SIAM Review)
An official website of the United States government

Full Text Available