skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Data for Printability Prediction in Projection Two-Photon Lithography via Machine Learning Based Surrogate Modeling of Photopolymerization
{"Abstract":["Data description \nThis dataset presents the raw and augmented data that were used to train the machine learning (ML) models for classification of printing outcome in projection two-photon lithography (P-TPL). P-TPL is an additive manufacturing technique for the fabrication of cm-scale complex 3D structures with features smaller than 200 nm. The P-TPL process is further described in this article: \u201cSaha, S. K., Wang, D., Nguyen, V. H., Chang, Y., Oakdale, J. S., and Chen, S.-C., 2019, "Scalable submicrometer additive manufacturing," Science, 366(6461), pp. 105-109.\u201d This specific dataset refers to the case wherein a set of five line features were projected and the printing outcome was classified into three classes: \u2018no printing\u2019, \u2018printing\u2019, \u2018overprinting\u2019. \n \nEach datapoint comprises a set of ten inputs (i.e., attributes) and one output (i.e., target) corresponding to these inputs. The inputs are: optical power (P), polymerization rate constant at the beginning of polymer conversion (kp-0), radical quenching rate constant (kq), termination rate constant at the beginning of polymer conversion (kt-0), number of optical pulses, (N), kp exponential function shape parameter (A), kt exponential function shape parameter (B), quantum yield of photoinitiator (QY), initial photoinitiator concentration (PIo), and the threshold degree of conversion (DOCth). The output variable is \u2018Class\u2019 which can take these three values: -1 for the class \u2018no printing\u2019, 0 for the class \u2018printing\u2019, and 1 for the class \u2018overprinting\u2019. \n\nThe raw data (i.e., the non-augmented data) refers to the data generated from finite element simulations of P-TPL. The augmented data was obtained from the raw data by (1) changing the DOCth and re-processing a solved finite element model or (2) by applying physics-based prior process knowledge. For example, it is known that if a given set of parameters failed to print, then decreasing the parameters that are positively correlated with printing (e.g. kp-0, power), while keeping the other parameters constant would also lead to no printing. Here, positive correlation means that individually increasing the input parameter will lead to an increase in the amount of printing. Similarly, increasing the parameters that are negatively correlated with printing (e.g. kq, kt-0), while keeping the other parameters constant would also lead to no printing. The converse is true for those datapoints that resulted in overprinting. \n\nThe 'Raw.csv' file contains the datapoints generated from finite element simulations, the 'Augmented.csv' file contains the datapoints generated via augmentation, and the 'Combined.csv' file contains the datapoints from both files. The ML models were trained on the combined dataset that included both raw and augmented data."]}  more » « less
Award ID(s):
2045147
PAR ID:
10435731
Author(s) / Creator(s):
Publisher / Repository:
Mendeley
Date Published:
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Two-photon lithography (TPL) is a laser-based additive manufacturing technique that enables the printing of arbitrarily complex cm-scale polymeric 3D structures with sub-micron features. Although various approaches have been investigated to enable the printing of fine features in TPL, it is still challenging to achieve rapid sub-100 nm 3D printing. A key limitation is that the physical phenomena that govern the theoretical and practical limits of the minimum feature size are not well known. Here, we investigate these limits in the projection TPL (P-PTL) process, which is a high-throughput variant of TPL, wherein entire 2D layers are printed at once. We quantify the effects of the projected feature size, optical power, exposure time, and photoinitiator concentration on the printed feature size through finite element modeling of photopolymerization. Simulations are performed rapidly over a vast parameter set exceeding 10,000 combinations through a dynamic programming scheme, which is implemented on high-performance computing resources. We demonstrate that there is no physics-based limit to the minimum feature sizes achievable with a precise and well-calibrated P-TPL system, despite the discrete nature of illumination. However, the practically achievable minimum feature size is limited by the increased sensitivity of the degree of polymer conversion to the processing parameters in the sub-100 nm regime. The insights generated here can serve as a roadmap towards fast, precise, and predictable sub-100 nm 3D printing. 
    more » « less
  2. {"Abstract":["This is an extracted data product for radar bed reflectivity from Whillans Ice Plain, West Antarctica. The original data are hosted by the Center for Remote Sensing of Ice Sheets (CReSIS; see associated citation below). The files here can be recalculate and are meant to be used within a set of computational notebooks here:https://doi.org/10.5281/zenodo.10859135\n\nThere are two csv files included here, each structured as a Pandas dataframe. You can load them in Python like:df = pd.read_csv('./Picked_Bed_Power.csv')\n\nThe first file, 'Picked_Bed_Power.csv' is the raw, uncorrected power from the radar image at the bed pick provided by CReSIS. There are also other useful variables for georeferencing, flight attributes, etc.\n\nThe second file, 'Processed_Reflectivity.csv' is processed from the first file. Processing includes: 1) a spreading correction; 2) an attenuation correction; and, 3) a power adjustment flight days based on compared power at crossover points. This file also has identifiers for regions including "grounded ice", "ungrounded ice", and "subglacial lakes"."]} 
    more » « less
  3. {"Abstract":["This data set contains all classifications that the Gravity Spy Machine Learning model for LIGO glitches from the first three observing runs (O1, O2 and O3, where O3 is split into O3a and O3b). Gravity Spy classified all noise events identified by the Omicron trigger pipeline in which Omicron identified that the signal-to-noise ratio was above 7.5 and the peak frequency of the noise event was between 10 Hz and 2048 Hz. To classify noise events, Gravity Spy made Omega scans of every glitch consisting of 4 different durations, which helps capture the morphology of noise events that are both short and long in duration.<\/p>\n\nThere are 22 classes used for O1 and O2 data (including No_Glitch and None_of_the_Above), while there are two additional classes used to classify O3 data.<\/p>\n\nFor O1 and O2, the glitch classes were: 1080Lines, 1400Ripples, Air_Compressor, Blip, Chirp, Extremely_Loud, Helix, Koi_Fish, Light_Modulation, Low_Frequency_Burst, Low_Frequency_Lines, No_Glitch, None_of_the_Above, Paired_Doves, Power_Line, Repeating_Blips, Scattered_Light, Scratchy, Tomte, Violin_Mode, Wandering_Line, Whistle<\/p>\n\nFor O3, the glitch classes were: 1080Lines, 1400Ripples, Air_Compressor, Blip, Blip_Low_Frequency<\/strong>, Chirp, Extremely_Loud, Fast_Scattering<\/strong>, Helix, Koi_Fish, Light_Modulation, Low_Frequency_Burst, Low_Frequency_Lines, No_Glitch, None_of_the_Above, Paired_Doves, Power_Line, Repeating_Blips, Scattered_Light, Scratchy, Tomte, Violin_Mode, Wandering_Line, Whistle<\/p>\n\nIf you would like to download the Omega scans associated with each glitch, then you can use the gravitational-wave data-analysis tool GWpy. If you would like to use this tool, please install anaconda if you have not already and create a virtual environment using the following command<\/p>\n\n```conda create --name gravityspy-py38 -c conda-forge python=3.8 gwpy pandas psycopg2 sqlalchemy```<\/p>\n\nAfter downloading one of the CSV files for a specific era and interferometer, please run the following Python script if you would like to download the data associated with the metadata in the CSV file. We recommend not trying to download too many images at one time. For example, the script below will read data on Hanford glitches from O2 that were classified by Gravity Spy and filter for only glitches that were labelled as Blips with 90% confidence or higher, and then download the first 4 rows of the filtered table.<\/p>\n\n```<\/p>\n\nfrom gwpy.table import GravitySpyTable<\/p>\n\nH1_O2 = GravitySpyTable.read('H1_O2.csv')<\/p>\n\nH1_O2[(H1_O2["ml_label"] == "Blip") & (H1_O2["ml_confidence"] > 0.9)]<\/p>\n\nH1_O2[0:4].download(nproc=1)<\/p>\n\n```<\/p>\n\nEach of the columns in the CSV files are taken from various different inputs: <\/p>\n\n[\u2018event_time\u2019, \u2018ifo\u2019, \u2018peak_time\u2019, \u2018peak_time_ns\u2019, \u2018start_time\u2019, \u2018start_time_ns\u2019, \u2018duration\u2019, \u2018peak_frequency\u2019, \u2018central_freq\u2019, \u2018bandwidth\u2019, \u2018channel\u2019, \u2018amplitude\u2019, \u2018snr\u2019, \u2018q_value\u2019] contain metadata about the signal from the Omicron pipeline. <\/p>\n\n[\u2018gravityspy_id\u2019] is the unique identifier for each glitch in the dataset. <\/p>\n\n[\u20181400Ripples\u2019, \u20181080Lines\u2019, \u2018Air_Compressor\u2019, \u2018Blip\u2019, \u2018Chirp\u2019, \u2018Extremely_Loud\u2019, \u2018Helix\u2019, \u2018Koi_Fish\u2019, \u2018Light_Modulation\u2019, \u2018Low_Frequency_Burst\u2019, \u2018Low_Frequency_Lines\u2019, \u2018No_Glitch\u2019, \u2018None_of_the_Above\u2019, \u2018Paired_Doves\u2019, \u2018Power_Line\u2019, \u2018Repeating_Blips\u2019, \u2018Scattered_Light\u2019, \u2018Scratchy\u2019, \u2018Tomte\u2019, \u2018Violin_Mode\u2019, \u2018Wandering_Line\u2019, \u2018Whistle\u2019] contain the machine learning confidence for a glitch being in a particular Gravity Spy class (the confidence in all these columns should sum to unity). <\/p>\n\n[\u2018ml_label\u2019, \u2018ml_confidence\u2019] provide the machine-learning predicted label for each glitch, and the machine learning confidence in its classification. <\/p>\n\n[\u2018url1\u2019, \u2018url2\u2019, \u2018url3\u2019, \u2018url4\u2019] are the links to the publicly-available Omega scans for each glitch. \u2018url1\u2019 shows the glitch for a duration of 0.5 seconds, \u2018url2\u2019 for 1 seconds, \u2018url3\u2019 for 2 seconds, and \u2018url4\u2019 for 4 seconds.<\/p>\n\n```<\/p>\n\nFor the most recently uploaded training set used in Gravity Spy machine learning algorithms, please see Gravity Spy Training Set on Zenodo.<\/p>\n\nFor detailed information on the training set used for the original Gravity Spy machine learning paper, please see Machine learning for Gravity Spy: Glitch classification and dataset on Zenodo. <\/p>"]} 
    more » « less
  4. Abstract Two-photon lithography (TPL) is a direct laser writing process that enables the fabrication of cm-scale complex three-dimensional polymeric structures with submicrometer resolution. In contrast to the slow and serial writing scheme of conventional TPL, projection TPL (P-TPL) enables rapid printing of entire layers at once. However, process prediction remains a significant challenge in P-TPL due to the lack of computationally efficient models. In this work, we present machine learning-based surrogate models to predict the outcomes of P-TPL to >98% of the accuracy of a physics-based reaction-diffusion finite element simulation. A classification neural network was trained using data generated from the physics-based simulations. This enabled us to achieve computationally efficient and accurate prediction of whether a set of printing conditions will result in precise and controllable polymerization and the desired printing versus no printing or runaway polymerization. We interrogate this surrogate model to investigate the parameter regimes that are promising for successful printing. We predict combinations of photoresist reaction rate constants that are necessary to print for a given set of processing conditions, thereby generating a set of printability maps. The surrogate models reduced the computational time that is required to generate these maps from more than 10 months to less than a second. Thus, these models can enable rapid and informed selection of photoresists and printing parameters during process control and optimization. 
    more » « less
  5. We introduce the UConn Bubbles with Swatches dataset. This dataset contains images of voting bubbles, scanned from Connecticut ballots, either captured as grayscale (8 bpp) or color (RGB, 24 bpp) artifacts, and extracted through segmentation using ballot geometry. These images are organized into 4 groups of datasets. The stored file contains all data together in color and we manually convert to greyscale. Each image of a bubble is 40x50 pixels. The labels are produced from an optical lens scanner.  The first dataset, Gray-B (Bubbles), uses 42,679 images (40x50, 8 bpp) with blank (35,429 images) and filled (7,250 images) bubbles filled in by humans, but no marginal marks. There are two classes, mark and nonmark. The second dataset, RGB-B, is a 24 bpp color (RGB) version of Bubbles-Gray.  The third dataset, Gray-C (Combined), augments Gray-B with a collection of marginal marks called “swatches”, which are synthetic images that vary the position of signal to create samples close to the boundary of an optical lens scanner. The 423,703 randomly generated swatches place equal amounts of random noise throughout each image such that the amount of light is the same. This yields 466,382 labeled images. The fourth dataset, RGB-C, is a 24bpp color (RGB) version of Gray-C. The empty bubbles are bubbles that were printed by a commercial vendor. They have undergone registration and segmentation using predetermined coordinates. Marks are on paper printed by the same vendor. These datasets can be used for classification training. The .h5 has many levels of datasets as shown below.  The main dataset used for training is positional.  This is only separated into blank (non-mark) and vote (mark).  Whether the example is a bubble or a swatch is indicated by batch number.  See https://github.com/VoterCenter/Busting-the-Ballot/blob/main/Utilities/LoadVoterData.py for code that creates torch arrays for RGB-B and RGB-C. See the linked Github repo (https://github.com/VoterCenter/Busting-the-Ballot/blob/main/Utilities/VoterLab_Classifier_Functions.py) for grayscale conversion functions and other utilities.   Dataset structure: COLOR - POSITIONAL - INFORMATION / / / B/V/Q B/V/Q COLOR/POSITIONAL / / / IMAGE IMAGE B/V/Q / BACKGROUND RGB VALUES Images divided into 'batches' not all of which have dataInformation contains labels for all images. Q is the swatch data, while B and V are non-mark and mark respectively. 
    more » « less