Fiber bundles have become widely adopted for use in endoscopy, live-organism imaging, and other imaging applications. An inherent consequence of imaging with these bundles is the introduction of a honeycomb-like artifact that arises from the inter-fiber spacing, which obscures features of objects in the image. This artifact subsequently limits applicability and can make interpretation of the image-based data difficult. This work presents a method to reduce this artifact by on-axis rotation of the fiber bundle. Fiber bundle images were first low-pass and median filtered to improve image quality. Consecutive filtered images with rotated samples were then co-registered and averaged to generate a final, reconstructed image. The results demonstrate removal of the artifacts, in addition to increased signal contrast and signal-to-noise ratio. This approach combines digital filtering and spatial resampling to reconstruct higher-quality images, enhancing the utility of images acquired using fiber bundles.
more »
« less
Testing the LSST Difference Image Analysis Pipeline Using Synthetic Source Injection Analysis
Abstract We evaluate the performance of the Legacy Survey of Space and Time Science Pipelines Difference Image Analysis (DIA) on simulated images. By adding synthetic sources to galaxies on images, we trace the recovery of injected synthetic sources to evaluate the pipeline on images from the Dark Energy Science Collaboration Data Challenge 2. The pipeline performs well, with efficiency and flux accuracy consistent with the signal-to-noise ratio of the input images. We explore different spatial degrees of freedom for the Alard–Lupton polynomial-Gaussian image subtraction kernel and analyze for trade-offs in efficiency versus artifact rate. Increasing the kernel spatial degrees of freedom reduces the artifact rate without loss of efficiency. The flux measurements with different kernel spatial degrees of freedom are consistent. We also here provide a set of DIA flags that substantially filter out artifacts from the DIA source table. We explore the morphology and possible origins of the observed remaining subtraction artifacts and suggest that given the complexity of these artifact origins, a convolution kernel with a set of flexible bases with spatial variation may be needed to yield further improvements.
more »
« less
- Award ID(s):
- 2239364
- PAR ID:
- 10530505
- Publisher / Repository:
- IOP
- Date Published:
- Journal Name:
- The Astrophysical Journal
- Volume:
- 967
- Issue:
- 1
- ISSN:
- 0004-637X
- Page Range / eLocation ID:
- 10
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Abstract Image subtraction is essential for transient detection in time-domain astronomy. The point-spread function (PSF), photometric scaling, and sky background generally vary with time and across the field of view for imaging data taken with ground-based optical telescopes. Image subtraction algorithms need to match these variations for the detection of flux variability. An algorithm that can be fully parallelized is highly desirable for future time-domain surveys. Here we introduce the saccadic fast Fourier transform (SFFT) algorithm we developed for image differencing. SFFT uses aδ-function basis for kernel decomposition, and the image subtraction is performed in Fourier space. This brings about a remarkable improvement in computational performance of about an order of magnitude compared to other published image subtraction codes. SFFT can accommodate the spatial variations in wide-field imaging data, including PSF, photometric scaling, and sky background. However, the flexibility of theδ-function basis may also make it more prone to overfitting. The algorithm has been tested extensively on real astronomical data taken by a variety of telescopes. Moreover, the SFFT code allows for the spatial variations of the PSF and sky background to be fitted by spline functions.more » « less
-
he Compressive Sensing (CS) framework has demonstrated improved acquisition efficiency on a variety of clinical applications. Of interest to this work is Reflectance Confocal Microscopy (RCM), where CS can influence a drastic reduction in instrumentation complexity and image acquisition times. However, CS introduces the disadvantage of requiring a time consuming and computationally intensive process for image recovery. To mitigate this, the current document details our preliminary work on expanding a Deep-Learning architecture for the acquisition and fast recovery of RCM images using CS. We show preliminary recoveries of RCM images of both a synthetic target and heterogeneous skin tissue using a state-of-the-art network architecture from compressive measurements at various undersampling rates. In addition, we propose an application-specific addition to an established network architecture, and evaluate its ability to further increase the accuracy of recovered CS RCM images and remove visual artifacts. Our initial results show that it is possible to recover compressively sampled images at near-real time rates with comparable quality to established computationally intensive and time-consuming optimization-based methods common in CS applicationsmore » « less
-
Underwater image restoration aims to recover color, contrast, and appearance in underwater scenes, crucial for fields like marine ecology and archaeology. While pixel-domain diffusion methods work for simple scenes, they are computationally heavy and produce artifacts in complex, depth-varying scenes. We present a single-step latent diffusion method, SLURPP (Single-step Latent Underwater Restoration with Pretrained Priors), that overcomes these limitations by combining a novel network architecture with an accurate synthetic data generation pipeline. SLURPP combines pretrained latent diffusion models - which encode strong priors on the geometry and depth of scenes with an explicit scene decomposition, which allows one to model and account for the effects of light attenuation and backscattering. To train SLURPP, we design a physics-based underwater image synthesis pipeline that applies varied and realistic underwater degradation effects to existing terrestrial image datasets. We evaluate our method extensively on both synthetic and real-world benchmarks and demonstrate state-of-the-art performance.more » « less
-
Abstract Oral history indicates that a large wooden trough held in storage at the University of Kentucky’s William S. Webb Museum of Anthropology was a component of the saltpeter mining operation in Mammoth Cave in the late 18th and early 19th centuries, worked largely by enslaved persons. We used multiple heritage science methods, including radiocarbon wiggle-match dating, tree-ring dating, scanning electron microscopy-energy dispersive X-ray spectroscopy (SEM–EDS), and optical scanning, combined with historical research, to examine the trough. Our analysis supports the oral history of the trough as an artifact of the mining system in Mammoth Cave. This case study illustrates how heritage science methods can provide corroboration for the origins and biographies of poorly documented historical artifacts.more » « less
An official website of the United States government

