SCI Publications
2025
T. Patel, T.A.J. Ouermi, T. Athawale,, C.R. Johnson.
Fast HARDI Uncertainty Quantification and Visualization with Spherical Sampling, In Computer Graphics Forum, Vol. 44, No. 3, pp. 1--12. 2025.
In this paper, we study uncertainty quantification and visualization of orientation distribution functions (ODF), which corresponds to the diffusion profile of high angular resolution diffusion imaging (HARDI) data. The shape inclusion probability (SIP) function is the state-of-the-art method for capturing the uncertainty of ODF ensembles. The current method of computing the SIP function with a volumetric basis exhibits high computational and memory costs, which can be a bottleneck to integrating uncertainty into HARDI visualization techniques and tools. We propose a novel spherical sampling framework for faster computation of the SIP function with lower memory usage and increased accuracy. In particular, we propose direct extraction of SIP isosurfaces, which represent confidence intervals indicating spatial uncertainty of HARDI glyphs, by performing spherical sampling of ODFs. Our spherical sampling approach requires much less sampling than the state-of-the-art volume sampling method, thus providing significantly enhanced performance, scalability, and the ability to perform implicit ray tracing. Our experiments demonstrate that the SIP isosurfaces extracted with our spherical sampling approach can achieve up to 8164× speedup, 37282× memory reduction, and 50.2% less SIP isosurface error compared to the classical volume sampling approach. We demonstrate the efficacy of our methods through experiments on synthetic and human-brain HARDI datasets.
A. C. Peterson, M. R. Requist, J. C. Benna, J. R. Nelson, S. Elhabian, C. de Cesar Netto, T. C. Beals, A. L. Lenz.
Talar Morphology of Charcot-Marie-Tooth Patients With Cavovarus Feet, In Foot & Ankle International, Sage Publications, 2025.
DOI: 10.1177/10711007241309915
PubMed ID: 39937093
Background:
Methods:
Results:
Conclusion:
P. Ramonetti, M. Floca, K. O'Laughlin, A. Gupta, M. Parashar, I. Altintas.
National Data Platform's Education Hub, Subtitled arXiv:2510.12820v1, 2025.
As demand for AI literacy and data science education grows, there is a critical need for infrastructure that bridges the gap between research data, computational resources, and educational experiences. To address this gap, we developed a first-of-its-kind Education Hub within the National Data Platform. This hub enables seamless connections between collaborative research workspaces, classroom environments, and data challenge settings. Early use cases demonstrate the effectiveness of the platform in supporting complex and resource-intensive educational activities. Ongoing efforts aim to enhance the user experience and expand adoption by educators and learners alike.
A. Roberts, J. Marquez, K.H. NG, K. Mickelson, A. Panta, G. Scorzelli, A. Gooch, P. Cushman, M. Fritts, H. Neog, V. Pascucci, M. Taufer.
The Making of a Community Dark Matter Dataset with the National Science Data Fabric, Subtitled arXiv:2507.13297v1, 2025.
Dark matter is believed to constitute approximately 85% of the universe’s matter, yet its fundamental nature remains elusive. Direct detection experiments, though globally deployed, generate data that is often locked within custom formats and non-reproducible software stacks, limiting interdisciplinary analysis and innovation. This paper presents a collaboration between the National Science Data Fabric (NSDF) and dark matter researchers to improve accessibility, usability, and scientific value of a calibration dataset collected with Cryogenic Dark Matter Search (CDMS) detectors at the University of Minnesota. We describe how NSDF services were used to convert data from a proprietary format into an open, multi-resolution IDX structure; develop a web-based dashboard for easily viewing signals; and release a Python-compatible CLI to support scalable workflows and machine learning applications. These contributions enable broader use of high-value dark matter datasets, lower the barrier to entry for new collaborators, and support reproducible, cross-disciplinary research.
R. Basu Roy, T. Patel, B. Li, S. Samsi, V. Gadepally, D. Tiwari.
GreenMix: Energy-Efficient Serverless Computing via Randomized Sketching on Asymmetric Multi-Cores, In Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, Association for Computing Machinery, 2025.
ISBN: 9798400714665
DOI: 10.1145/3712285.3759861
GreenMix is motivated by the renewed interest in asymmetric multi-core processors and the emergence of the serverless computing model. Asymmetric multi-cores offer better energy and performance trade-offs by placing different core types on the same die. However, existing serverless scheduling techniques do not leverage these benefits. GreenMix is the first serverless work to reduce energy and serverless keep-alive costs while meeting QoS targets by leveraging asymmetric multi-cores. GreenMix employs randomized sketching, tailored for serverless execution and keep-alive, to perform within 10% of the optimal solution in terms of energy efficiency and keep-alive cost reduction. GreenMix’s effectiveness is demonstrated through evaluations on clusters of ARM big.LITTLE and Intel Alder Lake asymmetric processors. It outperforms competing state-of-the-art schedulers, offering a novel approach for energy-efficient serverless computing.
S. Saha, S. Joshi, R. Whitaker.
ARD-VAE: A Statistical Formulation to Find the Relevant Latent Dimensions of Variational Autoencoders, Subtitled arXiv:2501.10901, 2025.
The variational autoencoder (VAE) is a popular, deep, latent-variable model (DLVM) due to its simple yet effective formulation for modeling the data distribution. Moreover, optimizing the VAE objective function is more manageable than other DLVMs. The bottleneck dimension of the VAE is a crucial design choice, and it has strong ramifications for the model’s performance, such as finding the hidden explanatory factors of a dataset using the representations learned by the VAE. However, the size of the latent dimension of the VAE is often treated as a hyperparameter estimated empirically through trial and error. To this end, we propose a statistical formulation to discover the relevant latent factors required for modeling a dataset. In this work, we use a hierarchical prior in the latent space that estimates the variance of the latent axes using the encoded data, which identifies the relevant latent dimensions. For this, we replace the fixed prior in the VAE objective function with a hierarchical prior, keeping the remainder of the formulation unchanged. We call the proposed method the automatic relevancy detection in the variational autoencoder (ARD-VAE). We demonstrate the efficacy of the ARD-VAE on multiple benchmark datasets in finding the relevant latent dimensions and their effect on different evaluation metrics, such as FID score and disentanglement analysis.
S. Saha, S. Joshi, R. Whitaker.
Disentanglement Analysis in Deep Latent Variable Models Matching Aggregate Posterior Distributions, Subtitled arXiv:2501.15705, 2025.
Deep latent variable models (DLVMs) are designed to learn meaningful representations in an unsupervised manner, such that the hidden explanatory factors are interpretable by independent latent variables (aka disentanglement). The variational autoencoder (VAE) is a popular DLVM widely studied in disentanglement analysis due to the modeling of the posterior distribution using a factorized Gaussian distribution that encourages the alignment of the latent factors with the latent axes. Several metrics have been proposed recently, assuming that the latent variables explaining the variation in data are aligned with the latent axes (cardinal directions). However, there are other DLVMs, such as the AAE and WAE-MMD (matching the aggregate posterior to the prior), where the latent variables might not be aligned with the latent axes. In this work, we propose a statistical method to evaluate disentanglement for any DLVMs in general. The proposed technique discovers the latent vectors representing the generative factors of a dataset that can be different from the cardinal latent axes. We empirically demonstrate the advantage of the method on two datasets.
S. Saha, R. Whitaker.
AdaSemSeg: An Adaptive Few-shot Semantic Segmentation of Seismic Facies, Subtitled arXiv:2501.16760, 2025.
Automated interpretation of seismic images using deep learning methods is challenging because of the limited availability of training data. Few-shot learning is a suitable learning paradigm in such scenarios due to its ability to adapt to a new task with limited supervision (small training budget). Existing few-shot semantic segmentation (FSSS) methods fix the number of target classes. Therefore, they do not support joint training on multiple datasets varying in the number of classes. In the context of the interpretation of seismic facies, fixing the number of target classes inhibits the generalization capability of a model trained on one facies dataset to another, which is likely to have a different number of facies. To address this shortcoming, we propose a few-shot semantic segmentation method for interpreting seismic facies that can adapt to the varying number of facies across the dataset, dubbed the AdaSemSeg. In general, the backbone network of FSSS methods is initialized with the statistics learned from the ImageNet dataset for better performance. The lack of such a huge annotated dataset for seismic images motivates using a self-supervised algorithm on seismic datasets to initialize the backbone network. We have trained the AdaSemSeg on three public seismic facies datasets with different numbers of facies and evaluated the proposed method on multiple metrics. The performance of the AdaSemSeg on unseen datasets (not used in training) is better than the prototype-based few-shot method and baselines.
A. Sahistan, S. Zellmann, N. Morrical, V. Pascucci, I. Wald.
Multi-Density Woodcock Tracking: Efficient & High-Quality Rendering for Multi-Channel Volumes, In Eurographics Symposium on Parallel Graphics and Visualization, Eurographics, 2025.
Volume rendering techniques for scientific visualization have increasingly transitioned toward Monte Carlo (MC) methods in recent years due to their flexibility and robustness. However, their application in multi-channel visualization remains underexplored. Traditional compositing-based approaches often employ arbitrary color blending functions, which lack a physical basis and can obscure data interpretation. We introduce multi-density Woodcock tracking, a simple and flexible extension of Woodcock tracking for multi-channel volume rendering that leverages the strengths of Monte Carlo methods to generate high-fidelity visuals. Our method offers a physically grounded solution for inter-channel color blending and eliminates the need for arbitrary blending functions. We also propose a unified blending modality by generalizing Woodcock’s distance tracking method, facilitating seamless integration of alternative blending functions from prior works. Through evaluation across diverse datasets, we demonstrate that our approach maintains real-time interactivity while achieving high-quality visuals by accumulating frames over time.
S.A. Sakin, K.E. Isaacs.
Managing Data for Scalable and Interactive Event Sequence Visualization, Subtitled arXiv:2508.03974, 2025.
Parallel event sequences, such as those collected in program execution traces and automated manufacturing pipelines, are typically visualized as interactive parallel timelines. As the dataset size grows, these charts frequently experience lag during common interactions such as zooming, panning, and filtering. Summarization approaches can improve interaction performance, but at the cost of accuracy in representation. To address this challenge, we introduce ESeMan (Event Sequence Manager), an event sequence management system designed to support interactive rendering of timeline visualizations with tunable accuracy. ESeMan employs hierarchical data structures and intelligent caching to provide visualizations with only the data necessary to generate accurate summarizations with significantly reduced data fetch time. We evaluate ESeMan's query times against summed area tables, M4 aggregation, and statistical sub-sampling on a variety of program execution traces. Our results demonstrate ESeMan provides better performance, achieving sub-100ms fetch times while maintaining visualization accuracy at the pixel level. We further present our benchmarking harness, enabling future performance evaluations for event sequence visualization.
A. Salinas, I. Sohail, V. Pascucci, P. Stefanakis, S. Amjad, A. Panta, R. Schigas, T. Chun-Yiu Chui, N. Duboc, M. Farrokhabadi, R. Stull.
Climate Data for Power Systems Applications: Lessons in Reusing Wildfire Smoke Data for Solar PV Studies, Subtitled arXiv:2509.09888v2, 2025.
Data reuse is using data for a purpose distinct from its original intent. As data sharing becomes more prevalent in science, enabling effective data reuse is increasingly important. In this paper, we present a power systems case study of data repurposing for enabling data reuse. We define data repurposing as the process of transforming data to fit a new research purpose. In our case study, we repurpose a geospatial wildfire smoke forecast dataset into a historical dataset. We analyze its efficacy toward analyzing wildfire smoke impact on solar photovoltaic energy production. We also provide documentation and interactive demos for using the repurposed dataset. We identify key enablers of data reuse including metadata standardization, contextual documentation, and communication between data creators and reusers. We also identify obstacles to data reuse such as risk of misinterpretation and barriers to efficient data access. Through an iterative approach to data repurposing, we demonstrate how leveraging and expanding knowledge transfer infrastructures like online documentation, interactive visualizations, and data streaming directly address these obstacles. The findings facilitate big data use from other domains for power systems applications and grid resiliency.
A. Samanta, Y. Jiang, R. Stutsman, R.B. Roy.
Not Just Fast, But Also Sustainable: Rethinking Network Routing, In MAIoT '25: Proceedings of the Middleware for Autonomous AIoT Systems in the Computing Continuum , ACM, 2025.
The carbon footprint of networking poses a significant environmental concern. In this paper, we show how transmission latency-focused traditional routing can be sub-optimal in terms of carbon footprint. Based on the dynamic factors that affect a network’s carbon footprint and transmission latency, we design a carbon-aware routing solution. We evaluate our solution on real backbone Internet topologies to show that there is an opportunity to minimize the network’s carbon emissions with only modest latency penalties.
A. Samanta, R. Stutsman, R.B. Roy.
GridGreen: Integrating Serverless Computing in HPC Systems for Performance and Sustainability, In SoCC '25: Proceedings of the 2025 ACM Symposium on Cloud Computing, ACM, 2025.
We present GridGreen, a scheduling framework that improves the sustainability and performance of scientific workflow execution by integrating serverless computing with traditional on-premise high performance computing (HPC) clusters. GridGreen allocates workflow components across HPC and serverless environments leveraging spatio-temporal variation of carbon intensity and component execution characteristics. It incorporates component-level optimization, speculative pre-warming, I/O-aware data management, and fallback adaptation to jointly minimize carbon footprint and service time under user-defined cost constraints. Our evaluations on large-scale bioinformatics workflows across leadership-class HPC facilities and cloud-based serverless regions demonstrate that GridGreen achieves robust, cost-effective execution while improving carbon efficiency.
A. Samanta, Y. Jiang, R. Stutsman, R.B. Roy.
Water Footprint of Datacenter Applications: Methodological Implications of Manufacturing, Operational, and Decommissioning Phases, In SoCC '25: Proceedings of the 2025 ACM Symposium on Cloud Computing , ACM, 2025.
Rising computational demands have made cloud datacenters’ water footprint a critical concern. We demonstrate how different water footprint accounting methodologies – incorporating operational, manufacturing, and decommissioning water consumption, impact measurements and highlight the need for methodology standardization for water-aware operations. Our analysis reveals opportunities for water-aware scheduling in datacenters by considering regional water variations and lifecycle impacts.
J. Schuchart, P. Diehl, M. Bauer, A. Bouteiller, G. Daiss, E. Kayraklioglu, S. Khandekar, T. Herault, J. K. Holmen, R. S. Rao, A. Strack, E. Schlaughter, J. C. Spinti, J. N. Thornock, A. Aiken, O. Aumage, M. Berzins, G. Bosilca, B. L. hamberlain, H. Kaiser,, L. Kale.
A Survey of Distributed Asynchronous Many-Task Models and Their Applications, Subtitled Authorea Preprints, 2025.
Asynchronous many-task (AMT) runtime systems have become an important paradigm for expressing fine-grained parallelism and managing asynchrony in high-performance computing (HPC). Originating from early dataflow concepts, AMTs have evolved to enable dynamic task generation, explicit dependency management, and asynchronous execution, facilitating the overlap of computation and communication. These capabilities address the limitations of traditional bulk-synchronous models, such as those employed in MPI+X, which can struggle with irregular, adaptive, or data-driven workloads. This survey provides a comprehensive overview of representative distributed AMT systems—including Charm++, HPX, Legion, PaRSEC, Uintah, Chapel, and StarPU—focusing on their design principles, execution models, and runtime mechanisms for scheduling, communication, and synchronization. We examine how these systems tackle key challenges such as load imbalance, runtime overheads, programmability, and performance portability. In addition, the paper discusses application domains where AMTs have demonstrated tangible benefits and highlights the conditions under which their use is most advantageous. The goal of this survey is to equip researchers and practitioners with a clear understanding of distributed AMT models and to provide guidance for selecting and applying the most suitable runtime system for specific computational objectives.
C. Scully-Allison, K. Williams, S. Brink, O. Pearce, K. Isaacs.
A Tale of Two Models: Understanding Data Workers' Internal and External Representations of Complex Data, Subtitled arXiv:2501.09862v2, 2025.
Data workers may have a different mental model of their data than the one reified in code. Understanding the organization of their data is necessary for analyzing data, be it through scripting, visualization, or abstract thought. More complicated organizations, such as tables with attached hierarchies, may tax people’s ability to think about and interact with data. To better understand and ultimately design for these situations, we conduct a study across a team of ten people working with the same reified data model. Through interviews and sketching, we probed their conception of the data model and developed themes through reflexive data analysis. Participants had diverse data models that differed from the reified data model, even among team members who had designed the model, resulting in parallel hazards limiting their ability to reason about the data. From these observations, we suggest potential design interventions for data analysis processes and tools.
C. Scully-Allison, K. Menear, K. Potter, A. McNutt, K.E. Isaacs, D. Duplyakin.
Same Data, Different Audiences: Using Personas to Scope a Supercomputing Job Queue Visualization, 2025.
Domain-specific visualizations sometimes focus on narrow, albeit important, tasks for one group of users. This focus limits the utility of a visualization to other groups working with the same data. While tasks elicited from other groups can present a design pitfall if not disambiguated, they also present a design opportunity -- development of visualizations that support multiple groups. This development choice presents a trade off of broadening the scope but limiting support for the more narrow tasks of any one group, which in some cases can enhance the overall utility of the visualization. We investigate this scenario through a design study where we develop \textitGuidepost, a notebook-embedded visualization of supercomputer queue data that helps scientists assess supercomputer queue wait times, machine learning researchers understand prediction accuracy, and system maintainers analyze usage trends. We adapt the use of personas for visualization design from existing literature in the HCI and software engineering domains and apply them in categorizing tasks based on their uniqueness across the stakeholder personas. Under this model, tasks shared between all groups should be supported by interactive visualizations and tasks unique to each group can be deferred to scripting with notebook-embedded visualization design. We evaluate our visualization with nine expert analysts organized into two groups: a "research analyst" group that uses supercomputer queue data in their research (representing the Machine Learning researchers and Jobs Data Analyst personas) and a "supercomputer user" group that uses this data conditionally (representing the HPC User persona). We find that our visualization serves our three stakeholder groups by enabling users to successfully execute shared tasks with point-and-click interaction while facilitating case-specific programmatic analysis workflows.
S. Sebbio, L. Carnevale, D. Balouek, M. Parashar, M. Villari.
Data-Driven Operational Artificial Intelligence for Computing Continuum: A Natural Disaster Management Use Case, In 2025 IEEE 25th International Symposium on Cluster, Cloud and Internet Computing Workshops (CCGridW), pp. 92-99. 2025.
The increasing frequency of documented natural disasters can be attributed to advances in communication technologies, such as satellites, the Internet, and smart devices that facilitate better disaster reporting. This is coupled with an actual rise in the occurrence of such events and improved documentation of their impacts. These trends underscore the pressing need for scalable and intelligent technological solutions to efficiently process large datasets, allowing informed decision-making and effective disaster response. This study presents a Computing Continuum framework that integrates intelligence across cloud, edge and deep edge tiers for efficient disaster data processing. A significant characteristic is the incorporation of Artificial Intelligence for IT Operations (AIOps), which leverages machine learning and analytics to facilitate dynamic resource management and adaptive system modeling, thereby addressing the intricate challenges posed by disaster scenarios. The architecture encompasses an AI-driven framework for monitoring and managing service, network, and infrastructure layers, tailoring policies to specific disaster needs. The proposed framework is applied to wildfire management, leveraging an AI Operation Manager to coordinate sensor-equipped drones for real-time data acquisition and processing. Operating at the deep edge tier, these drones transmit environmental data to edge and cloud infrastructures for analysis. This multi-tiered approach improves situational awareness, disaster response, and resource utilization.
M. Shao, S. Joshi.
Domain-Shift Immunity in Deep Deformable Registration via Local Feature Representations, Subtitled arXiv:2512.23142v1, 2025.
Deep learning has advanced deformable image registration, surpassing traditional optimization-based methods in both accuracy and efficiency. However, learning-based models are widely believed to be sensitive to domain shift, with robustness typically pursued through large and diverse training datasets, without explaining the underlying mechanisms. In this work, we show that domain-shift immunity is an inherent property of deep deformable registration models, arising from their reliance on local feature representations rather than global appearance for deformation estimation. To isolate and validate this mechanism, we introduce UniReg, a universal registration framework that decouples feature extraction from deformation estimation using fixed, pre-trained feature extractors and a UNet-based deformation network. Despite training on a single dataset, UniReg exhibits robust cross-domain and multi-modal performance comparable to optimization-based methods. Our analysis further reveals that failures of conventional CNN-based models under modality shift originate from dataset-induced biases in early convolutional layers. These findings identify local feature consistency as the key driver of robustness in learning-based deformable registration and motivate backbone designs that preserve domain-invariant local features.
H. Shrestha, J. Wilburn, B. Bollen, A.M. McNutt, A. Lex, L. Harrison.
ReVISitPy: Python Bindings for the reVISit Study Framework, In Eurovis 2025, 2025.
User experiments are an important part of visualization research, yet they remain costly, time-consuming to create, and difficult to prototype and pilot. The process of prototyping a study-from initial design to data collection and analysis-often requires the use of multiple systems (e.g. webservers and databases), adding complexity. We present reVISitPy, a Python library that enables visualization researchers to design, pilot deployments, and analyze pilot data entirely within a Jupyter notebook. Re- VISitPy provides a higher-level Python interface for the reVISit Domain-Specific Language (DSL) and study framework, which traditionally relies on manually authoring complex JSON configuration files. As study configurations grow larger, editing raw JSON becomes increasingly tedious and error-prone. By streamlining the configuration, testing, and preliminary analysis workflows, reVISitPy reduces the overhead of study prototyping and helps researchers quickly iterate on study designs before full deployment through the reVISit framework.
Page 5 of 149
