<?xml-model href='http://www.tei-c.org/release/xml/tei/custom/schema/relaxng/tei_all.rng' schematypens='http://relaxng.org/ns/structure/1.0'?><TEI xmlns="http://www.tei-c.org/ns/1.0">
	<teiHeader>
		<fileDesc>
			<titleStmt><title level='a'>RAPTOR: Reconfigurable Advanced Platform for Transdisciplinary Open Research</title></titleStmt>
			<publicationStmt>
				<publisher>ACM</publisher>
				<date>07/20/2025</date>
			</publicationStmt>
			<sourceDesc>
				<bibl> 
					<idno type="par_id">10632875</idno>
					<idno type="doi"></idno>
					
					<author>Hamed Najafi</author><author>Pratik Poudel</author><author>Kiavash Bahreini</author><author>Julio Ibarra</author><author>Fahad Saeed</author><author>Yuepeng Li</author><author>Jayantha Obeysekera</author><author>Jason Liu</author>
				</bibl>
			</sourceDesc>
		</fileDesc>
		<profileDesc>
			<abstract><ab><![CDATA[Scienti!c research is increasingly relying on complex work"ows that span multiple computing paradigms, including high-performance computing (HPC), high-throughput computing (HTC), and machine learning/arti!cial intelligence (ML/AI). Traditional monolithic computing infrastructures often struggle to accommodate these diverse and evolving demands. The Recon!gurable Advanced Platform for Transdisciplinary Open Research (RAPTOR) addresses this challenge by providing a dynamically recon!gurable computing environment that integrates with federated resources. RAPTOR's architecture enables dynamic provisioning between an HPC cluster and the Chameleon Cloud platform based on workload requirements, supporting bare-metal customization for specialized applications. This paper focuses on RAPTOR's recon!gurability features and demonstrates their e#ectiveness through quantitative performance evaluations across four scienti!c domains: computational proteomics, climate modeling, weather research, and hurricane risk assessment. Our results demonstrate that RAPTOR's recon!gurable design signi!cantly enhances research productivity by providing an appropriate computing environment for diverse computational needs.
CCS CONCEPTS• Computing methodologies → Distributed computing methodologies; • Computer systems organization → Distributed architectures; Recon!gurable computing.]]></ab></abstract>
		</profileDesc>
	</teiHeader>
	<text><body xmlns="http://www.tei-c.org/ns/1.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:xlink="http://www.w3.org/1999/xlink">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>Modern scienti!c research faces increasingly complex computational demands that span multiple computing paradigms. A single research project may require high-performance computing (HPC) for parallel simulations, high-throughput computing (HTC) for processing large volumes of independent tasks, and specialized environments for machine learning and arti!cial intelligence (ML/AI) applications. Traditional computing infrastructures-designed primarily for a single computing paradigm-often struggle to accommodate this diversity of needs, creating bottlenecks in scienti!c work"ows and hindering interdisciplinary collaboration.</p><p>We designed and developed the Recon!gurable Advanced Platform for Transdisciplinary Open Research (RAPTOR) to address some of these challenges through a novel architecture that dynamically adapts to diverse computational requirements. Unlike static and monolithic computing environments, RAPTOR implements a recon!gurable design that allows compute nodes to be provisioned on demand between an HPC cluster and the Chameleon Cloud platform <ref type="bibr">[8]</ref>. This "exibility enables researchers to access the most suitable computing environment for their speci!c applicationswhether that involves bare-metal provisioning for specialized software stacks, integration with the Open Science Grid (OSG) <ref type="bibr">[12]</ref> for high-throughput computing, or HPC cluster resources for tightly coupled parallel applications.</p><p>This paper demonstrates RAPTOR's e#ectiveness in supporting diverse scienti!c work"ows across multiple domains, with a speci!c focus on how its recon!gurability features address key computational challenges. We present performance benchmarks from four scienti!c applications-proteomics analysis, sea level prediction, weather research, and hurricane risk assessment-highlighting RAPTOR's speci!c advantages compared to traditional computing environments.</p><p>The remainder of this paper is organized as follows. Section 2 presents RAPTOR's design considerations in addressing the various computational demands for modern research. In Section 3, we discuss RAPTOR's architecture which provides recon!gurability and adaptability to a diverse range of scienti!c work"ows based on data characteristics and research requirements. In Section 4, we examine several representative applications to demonstrate RAP-TOR's capabilities in advancing scienti!c discovery. By outlining these aspects, this study highlights how RAPTOR enhances computational e$ciency, fosters interdisciplinary collaboration, and supports a broad spectrum of research domains. We conclude the paper and outline future directions in Section 5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">DESIGN CONSIDERATIONS AND POLYMORPHIC CAPABILITIES</head><p>Traditional computing infrastructures are typically optimized for a single computing paradigm, making them ine$cient for the increasingly diverse computational demands of modern scienti!c research. We identi!ed key challenges that informed RAPTOR's recon!gurable design:</p><p>&#8226; Resource allocation in"exibility: Fixed hardware allocations in traditional HPC environments lead to resource under-utilization for work"ows with varying computational needs over time. &#8226; Software environment constraints: Standard HPC clusters impose strict software environments that limit the deployment of specialized applications, such as emerging AI frameworks or domain-speci!c tools. &#8226; Interoperability barriers: Institutional boundaries restrict access to specialized computing resources, hindering collaborative research that could bene!t from shared infrastructure. &#8226; Work"ow transition overhead: Moving between computing paradigms (e.g., from HPC simulation to ML analysis) typically requires manual data transfer and environment recon!guration, creating work"ow bottlenecks. &#8226; Heterogeneous resources: Modern scienti!c applications increasingly require the seamless integration of diverse computing resources, including CPUs, GPUs, high-memory nodes, and storage systems, which would be challenging to achieve in any monolithic system environments.</p><p>To address these challenges, RAPTOR implements a recon!gurable architecture with !ve key polymorphic capabilities:</p><p>(1) High-performance computing: RAPTOR connects highend compute nodes to FIU's existing HPC cluster, providing a uni!ed environment for parallel applications with dynamic scaling capabilities that overcome !xed allocation limitations of traditional clusters. (2) High-throughput computing: Through OSG integration <ref type="bibr">[5]</ref>, RAPTOR enables the opportunistic execution of independent jobs across distributed resources, extending computational reach beyond institutional boundaries. (3) Data-intensive computing: RAPTOR nodes feature multiprocessor, multi-core servers with GPU acceleration, large memory (up to 1.5 TB/node), and scalable storage for big data processing with dynamic resource allocation. (4) Real-time on-demand provisioning: Using Chameleon's bare-metal provisioning, researchers gain full control over software stacks, enabling customized environments for applications with strict deadlines or specialized con!gurations. While container solutions like Kubernetes or Singularity provide portability on HPC and cloud platforms, RAPTOR's emphasis on bare-metal provisioning o#ers a complementary path that ensures unrestricted hardware access and allows full OS/kernel customization. (5) Federated resource integration: RAPTOR operates as a federated resource with Chameleon and OSG, enabling researchers to access and utilize resources based on their speci!c requirements while maintaining secure isolation.</p><p>These polymorphic capabilities are enabled through several key recon!gurability mechanisms: (1) network-level recon!guration using VLANs and software-de!ned networking, (2) system-level recon!guration with automated OS provisioning, (3) resource management integration across multiple scheduling systems, and (4) uni!ed storage access across computing environments. Together, these mechanisms allow RAPTOR to dynamically adapt its architecture to meet the diverse computational demands of scienti!c applications while maintaining e$cient resource utilization.</p><p>RAPTOR's design intentionally balances two key sets of tradeo#s. First, it weighs the increased management complexity and switching overhead of a recon!gurable system against the "exibility gains o#ered by specialized con!gurations. Second, it considers the security challenges introduced by federated access alongside the collaborative research opportunities that federation makes possible <ref type="bibr">[2,</ref><ref type="bibr">3]</ref>. Our evaluation of diverse scienti!c work"ows demonstrates that these trade-o#s are well-justi!ed by the signi!cant performance improvements and research productivity gains achieved through RAPTOR's recon!gurable architecture.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">RAPTOR ARCHITECTURE 3.1 System Architecture</head><p>RAPTOR's architecture is designed around the principle of recon-!gurability, enabling dynamic resource allocation across di#erent computing environments. As shown in Figure <ref type="figure">1</ref>, the system consists of three interconnected zones: the CHI Zone (Chameleon Cloud), the HPC Zone, and the Storage Zone <ref type="bibr">[1]</ref>.</p><p>The CHI Zone features RAPTOR Nodes with AMD EPYC processors, DDR4 memory, and NVIDIA A100 GPUs, provisioned via Chameleon's CHI-in-a-Box <ref type="bibr">[9]</ref>. It includes Management Nodes for lease management and a Storage Server providing 1 TB of scratch storage per user. RAPTOR's bare-metal access eliminates virtualization overhead for performance-sensitive applications.</p><p>The HPC Zone is connected to FIU's High-Performance Cluster <ref type="bibr">[4]</ref>, which features over 3,000 Intel cores, high-memory nodes, and 22 GPUs. The cluster comprises Master Nodes, compute nodes, a Login Node, and a Visualization Node, utilizing both Ethernet and In!niBand interconnects. It federates with OSG, extending capabilities beyond institutional boundaries, and uses DDN Storage for high-throughput data access.</p><p>A key innovation is the Provisioning Server which dynamically recon!gures nodes between zones using IPMI. This process is managed through a CLI tool that automates network settings, boot environments, and resource allocation. By default, nodes reside in the CHI Zone until speci!cally requested for HPC use.</p><p>The Storage Zone provides "exible, high-performance storage through a combination of technologies. The Storage Zone provides "exible storage through S3-compatible object storage, CEPH distributed storage <ref type="bibr">[15]</ref>, and iSCSI block storage, enabling researchers to scale resources according to speci!c work"ow requirements.</p><p>While the current system uses speci!c core zones, its architecture could theoretically support more. However, the scalability of managing numerous zones, particularly their control plane, requires future investigation.</p><p>The network infrastructure consists of two key switches that enable RAPTOR's recon!gurability. A Dell Switch is positioned between the CHI Zone and HPC Zone, facilitating the dynamic reassignment of computing resources between these environments. An Enterprise Switch connects the Storage Zone to both the HPC Zone and CHI Zone, ensuring that storage resources are accessible regardless of which computing environment is being used. These switches implement network-level recon!guration, enabling RAPTOR's dynamic resource allocation capabilities.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Recon!gurability</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>RAPTOR achieves dynamic recon!gurability through several key mechanisms:</head><p>Network-level recon!guration: The system uses Virtual Local Area Networks (VLANs) and software-de!ned networking <ref type="bibr">[10]</ref> to dynamically reassign nodes between the CHI and HPC environments. This recon!guration is managed by the Dell Switch and Enterprise Switch, which provide connectivity across zones.</p><p>System-level recon!guration: The Provisioning Server handles operating system deployment and con!guration, allowing nodes to boot into di#erent environments based on workload requirements. This includes specialized OS images for bare-metal deployments through Chameleon and standard HPC environments for cluster-based workloads.</p><p>Resource management integration: RAPTOR integrates with both the Slurm workload manager <ref type="bibr">[16]</ref> (for HPC workloads) and Chameleon's leasing system (for bare-metal provisioning). This allows researchers to request resources through familiar interfaces while leveraging RAPTOR's recon!gurability.</p><p>Automated work"ow support: The system includes tools for transitioning data and work"ows between environments, enabling researchers to move seamlessly between computing paradigms as their needs evolve. Importantly, RAPTOR's node recon!guration between zones (e.g., from the HPC Zone to the CHI Zone or vice versa) occurs between application executions. Running applications, including any tightly-coupled modules, are allowed to complete their execution on their allocated resources before these resources become eligible for reprovisioning into a di#erent environment for subsequent workloads. This operational model ensures that the recon!guration process itself does not impact the performance of active computations. Table <ref type="table">1</ref> summarizes RAPTOR's recon!gurability features and their bene!ts for scienti!c work"ows. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">SCIENTIFIC APPLICATIONS AND PERFORMANCE EVALUATION</head><p>To demonstrate RAPTOR's e#ectiveness in supporting diverse scienti!c work"ows, we evaluated its performance across four representative applications spanning di#erent computational paradigms. We highlight the speci!c recon!gurability features for each application that addressed key computational challenges.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">GPU-Accelerated Proteomics Analysis</head><p>Computational proteomics relies on comparing experimental mass spectrometry (MS) spectra against large databases of theoretical spectra to identify peptides. As database sizes grow to the terabyte scale for proteogenomic and metaproteomics analyses <ref type="bibr">[6]</ref>, traditional computing approaches become impractical, with processing times extending to weeks or months. RAPTOR's recon!gurability features enabled signi!cant performance improvements for the GiCOPS framework <ref type="bibr">[7]</ref>, which implements a distributed database search algorithm for peptide identi!cation. By leveraging bare-metal provisioning, researchers con!gured specialized environments with optimized CUDA and MPI installations-the ability to dynamically allocate GPU-accelerated nodes on demand provided "exibility unavailable on traditional HPC resources.</p><p>Without making any modi!cation to the code base, we recon!gured the RAPTOR environment (using bare-metal nodes), compiled MPI/CUDA GiCOPS code base, and then executed the experiments. Our experiments demonstrate that GiCOPS outperforms the existing GPU-based database search algorithms by more than 10 times in both closed-and open-search modes and 2 times as compared to the CPU-only HiCOPS. The code for GICOPS (MPI/CUDA) was compiled on RAPTOR without any modi!cations. The speedup attained when running GiCOPS on RAPTOR is comparable to the speedup obtained on traditional distributed memory architecture for same number of nodes.</p><p>Experiments used PRIDE datasets PXD055735, PXD055119, and PXD015384 <ref type="bibr">[13]</ref>, scaling MPI processes while maintaining constant parameters. Figure <ref type="figure">2</ref> shows execution time and speedup results. Our results show that we were able to successfully compile, troubleshoot, and execute MPI and CUDA based code bases with minimal modi!cation on the RAPTOR machines.</p><p>By scaling work"ows in the omics disciplines, RAPTOR empowers researchers to push the boundaries of computational biology, leading to faster biomarker discovery, improved disease modeling, and more e#ective drug development strategies. More speci!cally, our results show that RAPTOR's GPU-accelerated nodes provided a 6x-12x speedup (due to GPU node's availability) compared to CPU-only implementations (for 4 MPI processes with one GPU per process). This acceleration reduced processing time from days to hours, enabling more comprehensive analyses that would be impractical on traditional resources.</p><p>For this particular application, key bene!ts included customized CUDA/MPI con!gurations, on-demand GPU allocation, and seamless development-to-production transitions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Machine Learning for Sea Level Prediction</head><p>Coastal "ooding prediction requires processing large multivariate time series datasets to identify patterns and make forecasts. Traditional statistical approaches struggle with the complex, nonlinear relationships in these datasets, leading researchers to explore machine-learning techniques.</p><p>We used RAPTOR to evaluate PatchTST <ref type="bibr">[11]</ref>, a state-of-the-art multivariate time series transformer model, for predicting sea level variability at 21 National Oceanic and Atmospheric Administration (NOAA) tide gauges. The recon!gurable nature of RAPTOR allowed researchers to create custom environments with specialized deeplearning frameworks.</p><p>We compared training times for the PatchTST model on RAP-TOR with those on local workstations for a multivariate dataset comprising 21 variables across 52,696 time steps. The results show a consistent speedup on RAPTOR compared to local workstations. This acceleration enabled researchers to explore more hyperparameter con!gurations and longer prediction horizons, signi!cantly enhancing model accuracy for coastal "ooding prediction.</p><p>This research signi!cantly bene!ted from RAPTOR's recon!gurability, including: 1) a customized environment with specialized deep-learning frameworks; 2) having on-demand access to highmemory nodes for extensive dataset processing; and 3) the ability to scale computations based on model complexity. The Weather Research and Forecasting (WRF) model <ref type="bibr">[14]</ref> is a numerical weather prediction system designed for atmospheric research and operational forecasting. Running high-resolution WRF simulations requires substantial computational resources and specialized software con!gurations. RAPTOR's recon!gurability enabled researchers to deploy the WRF model for South Florida with customized environments optimized for meteorological simulations.</p><p>The bare-metal provisioning capability enabled the installation of specialized libraries and dependencies that are challenging to con!gure in traditional HPC environments. We evaluated WRF's performance on RAPTOR for high-resolution (1 km) simulations covering a 500 x 500 km domain centered on South Florida. Table <ref type="table">2</ref> compares di#erent simulation con!gurations.</p><p>The RAPTOR environment enabled high-resolution WRF simulations that were previously impractical due to software con!guration challenges. These simulations o#er crucial insights into local weather patterns that impact coastal communities, including severe weather events such as heavy rainfall and thunderstorms.</p><p>Key bene!ts for this application included specialized environment con!guration for meteorological simulations, access to highperformance I/O for multi-source data integration, and ability to adjust resource allocation based on simulation complexity.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Stochastic Storm Simulation for Hurricane Risk Assessment</head><p>The Florida Public Hurricane Loss Model (FPHLM) <ref type="bibr">[3]</ref> requires simulating over 140,000 synthetic storms to estimate property losses and insurance risks. This computationally intensive process involves independent simulations, which are ideal for high-throughput computing approaches. RAPTOR's integration with the Open Science Grid (OSG) enabled the e$cient execution of these simulations across distributed resources. By federating RAPTOR with OSG, researchers gained access to computational capacity far beyond local resources, significantly reducing the time required for comprehensive risk assessments.</p><p>For the 140,000 storm simulations, attempts to use only the local FIU High-Performance Cluster (RAPTOR's HPC Zone) faced significant throughput limitations. However, RAPTOR's orchestration of this workload across the federated Open Science Grid (OSG) dramatically reduced computation time from months to weeks. This highlights RAPTOR's cross-environment capability-leveraging OSG beyond the local HPC Zone-delivering quantitative performance gains compared to standard execution on institutional resources alone.</p><p>This research is bene!ted from RAPTOR's recon!gurability. More speci!cally, the federation with OSG enabled access to distributed computing resources beyond what has been o#ered at the university. Also, the high-throughput computing environment can be optimized for many independent simulations. E$cient data management allows for large simulation outputs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">DISCUSSION AND CONCLUSION</head><p>RAPTOR demonstrates that a recon!gurable computing platform can e#ectively address the diverse computational needs of modern scienti!c research. RAPTOR overcomes key limitations of traditional computing infrastructures by providing dynamic allocation between HPC and cloud environments, bare-metal provisioning for specialized applications, and federation with national resources. Our performance evaluations across four scienti!c domains show signi!cant advantages compared to traditional approaches:</p><p>&#8226; Accelerated computation: For GPU-accelerated proteomics work"ows, RAPTOR provided speedups of up to 12x. &#8226; Work"ow "exibility: The recon!gurable architecture enabled seamless transitions between computing paradigms, supporting complex work"ows that span HPC, HTC, and ML/AI applications.</p><p>&#8226; Resource e#ciency: Dynamic resource allocation optimizes utilization based on current research demands, avoiding the underutilization common in static computing environments. &#8226; Enhanced collaboration: The federation with national resources extended computational reach beyond institutional boundaries, fostering multi-institutional research collaborations that transcend geographical boundaries.</p><p>These bene!ts directly address the challenges identi!ed in the FlexScience workshop's focus on "exible computing infrastructures for scienti!c applications. RAPTOR's approach to recon!gurability o#ers valuable insights for designing the next generation of research computing platforms that can adapt to evolving computational demands.</p><p>Future work will integrate RAPTOR with our petabyte-scale storage system designed to support data-intensive applications. This integration will further enhance RAPTOR's capabilities for realtime simulations, high-throughput data processing, and machine learning-driven research.</p></div></body>
		</text>
</TEI>
