skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Title: Evolution of the Automatic Rhodopsin Modeling (ARM) Protocol
Abstract In recent years, photoactive proteins such as rhodopsins have become a common target for cutting-edge research in the field of optogenetics. Alongside wet-lab research, computational methods are also developing rapidly to provide the necessary tools to analyze and rationalize experimental results and, most of all, drive the design of novel systems. The Automatic Rhodopsin Modeling (ARM) protocol is focused on providing exactly the necessary computational tools to study rhodopsins, those being either natural or resulting from mutations. The code has evolved along the years to finally provide results that arereproducibleby any user,accurateandreliableso as to replicate experimental trends. Furthermore, the code isefficientin terms of necessary computing resources and time, andscalablein terms of both number of concurrent calculations as well as features. In this review, we will show how the code underlying ARM achieved each of these properties.  more » « less
Award ID(s):
1710191
PAR ID:
10375664
Author(s) / Creator(s):
; ; ; ;
Publisher / Repository:
Springer Science + Business Media
Date Published:
Journal Name:
Topics in Current Chemistry
Volume:
380
Issue:
3
ISSN:
2365-0869
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract Identifying impacted pathways is important because it provides insights into the biology underlying conditions beyond the detection of differentially expressed genes. Because of the importance of such analysis, more than 100 pathway analysis methods have been developed thus far. Despite the availability of many methods, it is challenging for biomedical researchers to learn and properly perform pathway analysis. First, the sheer number of methods makes it challenging to learn and choose the correct method for a given experiment. Second, computational methods require users to be savvy with coding syntax, and comfortable with command‐line environments, areas that are unfamiliar to most life scientists. Third, as learning tools and computational methods are typically implemented only for a few species (i.e., human and some model organisms), it is difficult to perform pathway analysis on other species that are not included in many of the current pathway analysis tools. Finally, existing pathway tools do not allow researchers to combine, compare, and contrast the results of different methods and experiments for both hypothesis testing and analysis purposes. To address these challenges, we developed an open‐source R package for Consensus Pathway Analysis (RCPA) that allows researchers to conveniently: (1) download and process data from NCBI GEO; (2) perform differential analysis using established techniques developed for both microarray and sequencing data; (3) perform both gene set enrichment, as well as topology‐based pathway analysis using different methods that seek to answer different research hypotheses; (4) combine methods and datasets to find consensus results; and (5) visualize analysis results and explore significantly impacted pathways across multiple analyses. This protocol provides many example code snippets with detailed explanations and supports the analysis of more than 1000 species, two pathway databases, three differential analysis techniques, eight pathway analysis tools, six meta‐analysis methods, and two consensus analysis techniques. The package is freely available on the CRAN repository. © 2024 The Authors. Current Protocols published by Wiley Periodicals LLC. Basic Protocol 1: Processing Affymetrix microarrays Basic Protocol 2: Processing Agilent microarrays Support Protocol: Processing RNA sequencing (RNA‐Seq) data Basic Protocol 3: Differential analysis of microarray data (Affymetrix and Agilent) Basic Protocol 4: Differential analysis of RNA‐Seq data Basic Protocol 5: Gene set enrichment analysis Basic Protocol 6: Topology‐based (TB) pathway analysis Basic Protocol 7: Data integration and visualization 
    more » « less
  2. We present a computational protocol for the fast and automated screening of excited-state hybrid quantum mechanics/molecular mechanics (QM/MM) models of rhodopsins to be used as fluorescent probes based on the automatic rhodopsin modeling protocol (a-ARM). Such “a-ARM fluorescence screening protocol” is implemented through a general Python-based driver, PyARM, that is also proposed here. The implementation and performance of the protocol are benchmarked using different sets of rhodopsin variants whose absorption and, more relevantly, emission spectra have been experimentally assessed. We show that, despite important limitations that make unsafe to use it as a black-box tool, the protocol reproduces the observed trends in fluorescence and it is capable of selecting novel potentially fluorescent rhodopsins. We also show that the protocol can be used in mechanistic investigations to discern fluorescence enhancement effects associated with a near degeneracy of the S1/S2 states or, alternatively, with a barrier generated via coupling of the S0/S1 wave functions. 
    more » « less
  3. Abstract MotivationVolumetric 3D object analyses are being applied in research fields such as structural bioinformatics, biophysics, and structural biology, with potential integration of artificial intelligence/machine learning (AI/ML) techniques. One such method, 3D Zernike moments, has proven valuable in analyzing protein structures (e.g., protein fold classification, protein–protein interaction analysis, and molecular dynamics simulations). Their compactness and efficiency make them amenable to large-scale analyses. Established methods for deriving 3D Zernike moments, however, can be inefficient, particularly when higher order terms are required, hindering broader applications. As the volume of experimental and computationally-predicted protein structure information continues to increase, structural biology has become a “big data” science requiring more efficient analysis tools. ResultsThis application note presents a Python-based software package, ZMPY3D, to accelerate computation of 3D Zernike moments by vectorizing the mathematical formulae and using graphical processing units (GPUs). The package offers popular GPU-supported libraries such as CuPy and TensorFlow together with NumPy implementations, aiming to improve computational efficiency, adaptability, and flexibility in future algorithm development. The ZMPY3D package can be installed via PyPI, and the source code is available from GitHub. Volumetric-based protein 3D structural similarity scores and transform matrix of superposition functionalities have both been implemented, creating a powerful computational tool that will allow the research community to amalgamate 3D Zernike moments with existing AI/ML tools, to advance research and education in protein structure bioinformatics. Availability and implementationZMPY3D, implemented in Python, is available on GitHub (https://github.com/tawssie/ZMPY3D) and PyPI, released under the GPL License. 
    more » « less
  4. Background:Early surgical exposure and research fellowships can influence medical students’ specialty choice, increase academic productivity, and impact residency match. However, to our knowledge, there is no published guidance on the programmatic evaluation and quality enhancement necessary for the sustainability of formal plastic surgery summer research programs for first year medical students. We present seven years (2013–2020) of institutional experience in an effort to inform program development at other institutions. Methods:From 2013 to 2016, a sole basic science research arm existed. In 2017, a clinical research arm was introduced, with several supplemental activities, including surgical skills curriculum. A formalized selection process was instituted in 2014. Participant feedback was analyzed annually. Long-term outcomes included continued research commitment, productivity, and residency match. Results:The applicant pool reached 96 applicants in 2019, with 85% from outside institutions. Acceptance rate reached 7% in 2020. With adherence to a scoring rubric for applicant evaluation, good to excellent interrater reliability was achieved (intraclass correlation coefficient = 0.75). Long-term outcomes showed that on average per year, 28% of participants continued involvement in departmental research and 29% returned for dedicated research. Upon finishing medical school, participants had a mean of 7 ± 4 peer-reviewed publications. In total, 62% of participants matched into a surgical residency program, with 54% in integrated plastic surgery. Conclusions:A research program designed for first year medical students interested in plastic surgery can achieve academic goals. Students are provided with mentorship, networking opportunities, and tools for self-guided learning and career development. 
    more » « less
  5. Computational tools have become increasingly important to the advancement of biological research. Despite the existence of data-sharing principles such as FAIR, there has been little attention paid to ensuring that computational tools provide a platform to compare experimental results and data across different studies. Tools often are lightly documented, non-extensible, and provide different levels of accuracy of their representation. This lack of standardization in biological research reduces the potential power for new insight and discovery and makes it hard for biologists to experiment, compare, and trust results between different studies. In this poster we present our experience using four tools that perform flux balance analysis, on a set of different metabolic models. We frame our work around a proposed principle, akin to FAIR, aimed at bioinformatics tools. We call this CORE (Comparable, Open, Reliable and Extensible), and find the biggest challenges to be with comparability between tools. We also find that while the tools are all open source, without a deep understanding of the code base, they have insufficient openness. We needed to reach out to developers to resolve many of our questions, and we were still left with unexpected behaviors. We present lessons learned as a path to future improvement using the CORE principles. 
    more » « less