skip to main content

Title: A Combinatorial Approach for Constructing Lattice Structures
Abstract Lattice structures exhibit unique properties including a large surface area and a highly distributed load-path. This makes them very effective in engineering applications where weight reduction, thermal dissipation, and energy absorption are critical. Furthermore, with the advent of additive manufacturing (AM), lattice structures are now easier to fabricate. However, due to inherent surface complexity, their geometric construction can pose significant challenges. A classic strategy for constructing lattice structures exploits analytic surface–surface intersection; this, however, lacks robustness and scalability. An alternate strategy is voxel mesh-based isosurface extraction. While this is robust and scalable, the surface quality is mesh-dependent, and the triangulation will require significant postdecimation. A third strategy relies on explicit geometric stitching where tessellated open cylinders are stitched together through a series of geometric operations. This was demonstrated to be efficient and scalable, requiring no postprocessing. However, it was limited to lattice structures with uniform beam radii. Furthermore, existing algorithms rely on explicit convex-hull construction which is known to be numerically unstable. In this paper, a combinatorial stitching strategy is proposed where tessellated open cylinders of arbitrary radii are stitched together using topological operations. The convex hull construction is handled through a simple and robust projection method, avoiding expensive more » exact-arithmetic calculations and improving the computational efficiency. This is demonstrated through several examples involving millions of triangles. On a typical eight-core desktop, the proposed algorithm can construct approximately up to a million cylinders per second. « less
; ;
Award ID(s):
Publication Date:
Journal Name:
Journal of Mechanical Design
Sponsoring Org:
National Science Foundation
More Like this
  1. An extensible open-source deterministic global optimizer (EAGO) programmed entirely in the Julia language is presented. EAGO was developed to serve the need for supporting higher-complexity user-defined functions (e.g. functions defined implicitly via algorithms) within optimization models. EAGO embeds a first-of-its-kind implementation of McCormick arithmetic in an Evaluator structure allowing for the construction of convex/concave relaxations using a combination of source code transformation, multiple dispatch, and context-specific approaches. Utilities are included to parse userdefined functions into a directed acyclic graph representation and perform symbolic transformations enabling dramatically improved solution speed. EAGO is compatible with a wide variety of local optimizers, the most exhaustive library of transcendental functions, and allows for easy accessibility through the JuMP modelling language. Together with Julia’s minimalist syntax and competitive speed, these powerful features make EAGO a versatile research platform enabling easy construction of novel meta-solvers, incorporation and utilization of new relaxations, and extension to advanced problem formulations encountered in engineering and operations research (e.g. multilevel problems, user-defined functions). The applicability and flexibility of this novel software is demonstrated on a diverse set of examples. Lastly, EAGO is demonstrated to perform comparably to state-of-the-art commercial optimizers on a benchmarking test set.
  2. Abstract

    Recent work has demonstrated that geometric deep learning methods such as graph neural networks (GNNs) are well suited to address a variety of reconstruction problems in high-energy particle physics. In particular, particle tracking data are naturally represented as a graph by identifying silicon tracker hits as nodes and particle trajectories as edges, given a set of hypothesized edges, edge-classifying GNNs identify those corresponding to real particle trajectories. In this work, we adapt the physics-motivated interaction network (IN) GNN toward the problem of particle tracking in pileup conditions similar to those expected at the high-luminosity Large Hadron Collider. Assuming idealized hit filtering at various particle momenta thresholds, we demonstrate the IN’s excellent edge-classification accuracy and tracking efficiency through a suite of measurements at each stage of GNN-based tracking: graph construction, edge classification, and track building. The proposed IN architecture is substantially smaller than previously studied GNN tracking architectures; this is particularly promising as a reduction in size is critical for enabling GNN-based tracking in constrained computing environments. Furthermore, the IN may be represented as either a set of explicit matrix operations or a message passing GNN. Efforts are underway to accelerate each representation via heterogeneous computing resources towards both high-levelmore »and low-latency triggering applications.

    « less
  3. Hydrodynamic performance of ships can be greatly improved by the formation of air cavities under ship bottom with the purpose to decrease water friction on the hull surface. The air-cavity ships using this type of drag reduction are usually designed for and typically effective only in a relatively narrow range of speeds and hull attitudes and sufficient rates of air supply to the cavity. To investigate the behavior of a small-scale air-cavity boat operating under both favorable and detrimental loading and speed conditions, a remotely controlled model hull was equipped with a data acquisition system, video camera and onboard sensors to measure air-cavity characteristics, air supply rate and the boat speed, thrust and trim in operations on open-water reservoirs. These measurements were captured by a data logger and also wirelessly transmitted to a ground station and video monitor. The experimental air-cavity boat was tested in a range of speeds corresponding to length Froude numbers between 0.17 and 0.5 under three loading conditions, resulting in near zero trim and significant bow-up and bow-down trim angles at rest. Reduced cavity size and significantly increased drag occurred when operating at higher speeds, especially in the bow-up trim condition. The other objective of thismore »study was to determine whether computational fluid dynamics simulations can adequately capture the recorded behavior of the boat and air cavity. A computational software Star-CCM+ was utilized with the VOF method employed for multi-phase flow, RANS approach for turbulence modeling, and economical mesh settings with refinements in the cavity region and near free surface. Upon conducting the mesh verification study, several experimental conditions were simulated, and approximate agreement with measured test data was found. Adaptive mesh refinement and time step controls were also applied to compare results with those obtained on the user-generated mesh. Adaptive controls improved resolution of complex shedding patterns from the air cavity but had little impact on overall results. The presented here experimental approach and obtained results indicate that both outdoor experimentation and computationally inexpensive modeling can be used in the process of developing air-cavity systems for ship hulls.« less
  4. Previous approaches on 3D shape segmentation mostly rely on heuristic processing and hand-tuned geometric descriptors. In this paper, we propose a novel 3D shape representation learning approach, Directionally Convolutional Network (DCN), to solve the shape segmentation problem. DCN extends convolution operations from images to the surface mesh of 3D shapes. With DCN, we learn effective shape representations from raw geometric features, i.e., face normals and distances, to achieve robust segmentation. More specifically, a two-stream segmentation framework is proposed: one stream is made up by the proposed DCN with the face normals as the input, and the other stream is implemented by a neural network with the face distance histogram as the input. The learned shape representations from the two streams are fused by an element-wise product. Finally, Conditional Random Field (CRF) is applied to optimize the segmentation. Through extensive experiments conducted on benchmark datasets, we demonstrate that our approach outperforms the current state-of-the-arts (both classic and deep learning-based) on a large variety of 3D shapes.
  5. Summary This paper is concerned with empirical likelihood inference on the population mean when the dimension $p$ and the sample size $n$ satisfy $p/n\rightarrow c\in [1,\infty)$. As shown in Tsao (2004), the empirical likelihood method fails with high probability when $p/n>1/2$ because the convex hull of the $n$ observations in $\mathbb{R}^p$ becomes too small to cover the true mean value. Moreover, when $p> n$, the sample covariance matrix becomes singular, and this results in the breakdown of the first sandwich approximation for the log empirical likelihood ratio. To deal with these two challenges, we propose a new strategy of adding two artificial data points to the observed data. We establish the asymptotic normality of the proposed empirical likelihood ratio test. The proposed test statistic does not involve the inverse of the sample covariance matrix. Furthermore, its form is explicit, so the test can easily be carried out with low computational cost. Our numerical comparison shows that the proposed test outperforms some existing tests for high-dimensional mean vectors in terms of power. We also illustrate the proposed procedure with an empirical analysis of stock data.