skip to main content

Search for: All records

Creators/Authors contains: "Zhao, Yang"

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. We present a double-yield-surface plasticity theory for transversely isotropic rocks that distinguishes between plastic deformation through the solid matrix and localized plasticity along the weak bedding planes. A recently developed anisotropic modified Cam-Clay model is adopted to model the plastic response of the solid matrix, while the Mohr-Coulomb friction law is used to represent the sliding mechanism along the weak bedding planes. For its numerical implementation, we derive an implicit return mapping algorithm for both the semi-plastic and fully plastic loading processes, as well as the corresponding algorithmic tangent operator for finite element problems. We validate the model with triaxial compression test data for three different transversely isotropic rocks and reproduce the undulatory variation of rock strength with bedding plane orientation. We also implement the proposed model in a finite element setting and investigate the deformation of rock surrounding a borehole subjected to fluid injection. We compare the results of simulations using the proposed double-yield-surface model with those generated using each single yield criterion to highlight the features of the proposed theory.
    Free, publicly-accessible full text available June 1, 2023
  2. We present a first-of-its-kind ultra-compact intelligent camera system, dubbed i-FlatCam, including a lensless camera with a computational (Comp.) chip. It highlights (1) a predict-then-focus eye tracking pipeline for boosted efficiency without compromising the accuracy, (2) a unified compression scheme for single-chip processing and improved frame rate per second (FPS), and (3) dedicated intra-channel reuse design for depth-wise convolutional layers (DW-CONV) to increase utilization. i-FlatCam demonstrates the first eye tracking pipeline with a lensless camera and achieves 3.16 degrees of accuracy, 253 FPS, 91.49 µJ/Frame, and 6.7mm×8.9mm×1.2mm camera form factor, paving the way for next-generation Augmented Reality (AR) and Virtual Reality (VR) devices.
    Free, publicly-accessible full text available June 12, 2023
  3. Eye tracking has become an essential human-machine interaction modality for providing immersive experience in numerous virtual and augmented reality (VR/AR) applications desiring high throughput (e.g., 240 FPS), small-form, and enhanced visual privacy. However, existing eye tracking systems are still limited by their: (1) large form-factor largely due to the adopted bulky lens-based cameras; (2) high communication cost required between the camera and backend processor; and (3) potentially concerned low visual privacy, thus prohibiting their more extensive applications. To this end, we propose, develop, and validate a lensless FlatCambased eye tracking algorithm and accelerator co-design framework dubbed EyeCoD to enable eye tracking systems with a much reduced form-factor and boosted system efficiency without sacrificing the tracking accuracy, paving the way for next-generation eye tracking solutions. On the system level, we advocate the use of lensless FlatCams instead of lens-based cameras to facilitate the small form-factor need in mobile eye tracking systems, which also leaves rooms for a dedicated sensing-processor co-design to reduce the required camera-processor communication latency. On the algorithm level, EyeCoD integrates a predict-then-focus pipeline that first predicts the region-of-interest (ROI) via segmentation and then only focuses on the ROI parts to estimate gaze directions, greatly reducing redundant computations andmore »data movements. On the hardware level, we further develop a dedicated accelerator that (1) integrates a novel workload orchestration between the aforementioned segmentation and gaze estimation models, (2) leverages intra-channel reuse opportunities for depth-wise layers, (3) utilizes input feature-wise partition to save activation memory size, and (4) develops a sequential-write-parallel-read input buffer to alleviate the bandwidth requirement for the activation global buffer. On-silicon measurement and extensive experiments validate that our EyeCoD consistently reduces both the communication and computation costs, leading to an overall system speedup of 10.95×, 3.21×, and 12.85× over general computing platforms including CPUs and GPUs, and a prior-art eye tracking processor called CIS-GEP, respectively, while maintaining the tracking accuracy. Codes are available at« less
    Free, publicly-accessible full text available June 11, 2023
  4. The record-breaking performance of deep neural networks (DNNs) comes with heavy parameter budgets, which leads to external dynamic random access memory (DRAM) for storage. The prohibitive energy of DRAM accesses makes it nontrivial for DNN deployment on resource-constrained devices, calling for minimizing the movements of weights and data in order to improve the energy efficiency. Driven by this critical bottleneck, we present SmartDeal, a hardware-friendly algorithm framework to trade higher-cost memory storage/access for lower-cost computation, in order to aggressively boost the storage and energy efficiency, for both DNN inference and training. The core technique of SmartDeal is a novel DNN weight matrix decomposition framework with respective structural constraints on each matrix factor, carefully crafted to unleash the hardware-aware efficiency potential. Specifically, we decompose each weight tensor as the product of a small basis matrix and a large structurally sparse coefficient matrix whose nonzero elements are readily quantized to the power-of-2. The resulting sparse and readily quantized DNNs enjoy greatly reduced energy consumption in data movement as well as weight storage, while incurring minimal overhead to recover the original weights thanks to the required sparse bit-operations and cost-favorable computations. Beyond inference, we take another leap to embrace energy-efficient training, by introducingmore »several customized techniques to address the unique roadblocks arising in training while preserving the SmartDeal structures. We also design a dedicated hardware accelerator to fully utilize the new weight structure to improve the real energy efficiency and latency performance. We conduct experiments on both vision and language tasks, with nine models, four datasets, and three settings (inference-only, adaptation, and fine-tuning). Our extensive results show that 1) being applied to inference, SmartDeal achieves up to 2.44x improvement in energy efficiency as evaluated using real hardware implementations and 2) being applied to training, SmartDeal can lead to 10.56x and 4.48x reduction in the storage and the training energy cost, respectively, with usually negligible accuracy loss, compared to state-of-the-art training baselines. Our source codes are available at:« less
    Free, publicly-accessible full text available March 2, 2023
  5. For many clinical applications, such as dermatology, optical coherence tomography (OCT) suffers from limited penetration depth due primarily to the highly scattering nature of biological tissues. Here, we present a novel implementation of dual-axis optical coherence tomography (DA-OCT) that offers improved depth penetration in skin imaging at 1.3 µm compared to conventional OCT. Several unique aspects of DA-OCT are examined here, including the requirements for scattering properties to realize the improvement and the limited depth of focus (DOF) inherent to the technique. To overcome this limitation, our approach uses a tunable lens to coordinate focal plane selection with image acquisition to create an enhanced DOF for DA-OCT. This improvement in penetration depth is quantified experimentally against conventional on-axis OCT using tissue phantoms and mouse skin. The results presented here suggest the potential use of DA-OCT in situations where a high degree of scattering limits depth penetration in OCT imaging.

  6. Free, publicly-accessible full text available January 1, 2023