Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to nonfederal websites. Their policies may differ from this site.

The ongoing electrification of the transportation fleet will increase the load on the electric power grid. Since both the transportation network and the power grid already experience periods of significant stress, joint analyses of both infrastructures will most likely be necessary to ensure acceptable operation in the future. To enable such analyses, this paper presents an open source testbed that jointly simulates highfidelity models of both the electric distribution system and the transportation network. The testbed utilizes two opensource simulators, OpenDSS to simulate the electric distribution system and the microscopic traffic simulator SUMO to simulate the traffic dynamics. Electric vehicle charging links the electric distribution system and the transportation network models at vehicle locations determined using publicly available parcel data. Leveraging highfidelity synthetic electric distribution system data from the SMARTDS project and transportation system data from OpenStreetMap, this testbed models the city of Greensboro, NC down to the household level. Moreover, the methodology and the supporting scripts released with the testbed allow adaption to other areas where highfidelity geolocated OpenDSS datasets are available. After describing the components and usage of the testbed, we exemplify applications enabled by the testbed via two scenarios modeling the extreme stresses encountered during evacuations.more » « lessFree, publiclyaccessible full text available May 1, 2025

Interval Markov decision processes are a class of Markov models where the transition probabilities between the states belong to intervals. In this paper, we study the problem of efficient estimation of the optimal policies in Interval Markov Decision Processes (IMDPs) with continuous action space. Given an IMDP, we show that the pessimistic (resp. the optimistic) value iterations, i.e., the value iterations under the assumption of a competitive adversary (resp. cooperative agent), are monotone dynamical systems and are contracting with respect to the infinitynorm. Inspired by this dynamical system viewpoint, we introduce another IMDP, called the actionspace relaxation IMDP. We show that the actionspace relaxation IMDP has two key features: (i) its optimal value is an upper bound for the optimal value of the original IMDP, and (ii) its value iterations can be efficiently solved using tools and techniques from convex optimization. We then consider the policy optimization problems at each step of the value iterations as a feedback controller of the value function. Using this system theoretic perspective, we propose an iterationdistributed imple mentation of the value iterations for approximating the optimal value of the actionspace relaxation IMDP.more » « lessFree, publiclyaccessible full text available December 15, 2024

In this paper, we present a toolbox for interval analysis in numpy, with an application to formal verification of neural network controlled systems. Using the notion of natural inclusion functions, we systematically construct interval bounds for a general class of mappings. The toolbox offers ef ficient computation of natural inclusion functions using compiled C code, as well as a familiar inter face in numpy with its canonical features, such as ndimensional arrays, matrix/vector operations, and vectorization. We then use this toolbox in for mal verification of dynamical systems with neural network controllers, through the composition of their inclusion functions.more » « less