SCI Publications
2025
W. Wu, M. Daemer, J.A. Weiss, A.M. Pouch, M.A. Jolley.
Novel point cloud registration approach for noninvasive patient specific estimation of leaflet strain from 3D images of heart valves, Subtitled arXiv:2510.06578, 2025.
Valvular heart disease is prevalent and a major contributor to heart failure. Valve leaflet strain is a promising metric for evaluating the mechanics underlying the initiation and progression of valvular pathology. However, robust and generalizable methods for noninvasively quantifying valvular strain from clinically acquired patient images remain limited. In this work, we present a novel feature-tracking framework for quantifying leaflet strain in atrioventricular valves using 3D echocardiographic images of pediatric and adult patients. Our method demonstrated superior accuracy in the assessment of anatomical deformation and strain of heart valves compared to other point-based approaches, as verified against a finite element benchmark. Further, our approach can robustly track inter-phase deformation of valves across highly variable morphologies without parameter tuning. Our analysis revealed that a median and interquartile range of the 1st principal strain greater than 0.5 is associated with leaflet billow (prolapse). Further investigation of the biomechanical signatures of heart valve disease has the potential to enhance prognostic assessment and longitudinal evaluation of valvular disease.
H. Xu, S.Y. Elhabian.
Adaptive Particle-Based Shape Modeling for Anatomical Surface Correspondence, Subtitled arXiv:2507.07379, 2025.
Particle-based shape modeling (PSM) is a family of approaches that automatically quantifies shape variability across anatomical cohorts by positioning particles (pseudo landmarks) on shape surfaces in a consistent configuration. Recent advances incorporate implicit radial basis function representations as self-supervised signals to better capture the complex geometric properties of anatomical structures. However, these methods still lack self-adaptivity—that is, the ability to automatically adjust particle configurations to local geometric features of each surface, which is essential for accurately representing complex anatomical variability. This paper introduces two mechanisms to increase surface adaptivity while maintaining consistent particle configurations: (1) a novel neighborhood correspondence loss to enable high adaptivity and (2) a geodesic correspondence algorithm that regularizes optimization to enforce geodesic neighborhood consistency. We evaluate the efficacy and scalability of our approach on challenging datasets, providing a detailed analysis of the adaptivity-correspondence trade-off and benchmarking against existing methods on surface representation accuracy and correspondence metrics.
S. Xu, M. Rasouli, R.M. Kirby, D. Moxey, H. Sundar.
A geometrically informed algebraic multigrid preconditioned iterative approach for solving high-order finite element systems, Subtitled arXiv:2512.15121v1, 2025.
Algebraic multigrid (AMG) is conventionally applied in a black-box fashion, agnostic to the underlying geometry. In this work, we propose that using geometric information -- when available -- to assist with setting up the AMG hierarchy is beneficial, especially for solving linear systems resulting from high-order finite element discretizations. High-order problems draw considerable interest to both the scientific and engineering communities, but lack efficient solvers, at least open-source codes, tailored for unstructured high-order discretizations targeting large-scale, real-world applications. For geometric multigrid, it is known that using p-coarsening before h-coarsening can provide better scalability, but setting up p-coarsening is non-trivial in AMG. We develop a geometrically informed algebraic multigrid (GIAMG) method, as well as an associated high-performance computing program, which is able to set up a grid hierarchy that includes p-coarsening at the top grids with minimal information of the geometry from the user. A major advantage of using p-coarsening with AMG -- beyond the benefits known in the context of geometric multigrid (GMG) -- is the increased sparsification of coarse grid operators. We extensively evaluate GIAMG by testing on the 3D Helmholtz and incompressible flow problems, and demonstrate mesh-independent convergence, and excellent parallel scalability. We also compare the performance of GIAMG with existing AMG packages, including Hypre and ML.
C. You, H. Dai, Y. Min, J.S. Sekhon, S. Joshi, J.S. Duncan.
The Silent Majority: Demystifying Memorization Effect in the Presence of Spurious Correlations, Subtitled arXiv:2501.00961v2, 2025.
Machine learning models often rely on simple spurious features – patterns in training data that correlate with targets but are not causally related to them, like image backgrounds in foreground classification. This reliance typically leads to imbalanced test performance across minority and majority groups. In this work, we take a closer look at the fundamental cause of such imbalanced performance through the lens of memorization, which refers to the ability to predict accurately on atypical examples (minority groups) in the training set but failing in achieving the same accuracy in the testing set. This paper systematically shows the ubiquitous existence of spurious features in a small set of neurons within the network, providing the first-ever evidence that memorization may contribute to imbalanced group performance. Through three experimental sources of converging empirical evidence, we find the property of a small subset of neurons or channels in memorizing minority group information. Inspired by these findings, we articulate the hypothesis: the imbalanced group performance is a byproduct of “noisy” spurious memorization confined to a small set of neurons. To further substantiate this hypothesis, we show that eliminating these unnecessary spurious memorization patterns via a novel framework during training can significantly affect the model performance on minority groups. Our experimental results across various architectures and benchmarks offer new insights on how neural networks encode core and spurious knowledge, laying the groundwork for future research in demystifying robustness to spurious correlation.
C. You, H. Dai, Y. Min, J.S. Sekhon, S. Joshi, J.S. Duncan.
Uncovering memorization effect in the presence of spurious correlations, In Nature Communications, 2025.
Machine learning models often rely on simple spurious features – patterns in training data that correlate with targets but are not causally related to them, like image backgrounds in foreground classification. This reliance typically leads to imbalanced test performance across minority and majority groups. In this work, we take a closer look at the fundamental cause of such imbalanced performance through the lens of memorization, which refers to the ability to predict accurately on atypical examples (minority groups) in the training set but failing in achieving the same accuracy in the testing set. This paper systematically shows the ubiquitous existence of spurious features in a small set of neurons within the network, providing the first-ever evidence that memorization may contribute to imbalanced group performance. Through three experimental sources of converging empirical evidence, we find the property of a small subset of neurons or channels in memorizing minority group information. Inspired by these findings, we hypothesize that spurious memorization, concentrated within a small subset of neurons, plays a key role in driving imbalanced group performance. To further substantiate this hypothesis, we show that eliminating these unnecessary spurious memorization patterns via a novel framework during training can significantly affect the model performance on minority groups. Our experimental results across various architectures and benchmarks offer new insights on how neural networks encode core and spurious knowledge, laying the groundwork for future research in demystifying robustness to spurious correlation.
Y. Zhang, M. A. Urquijo, R. G. Zitnay, K. Marks, R. L. Belote, M.M.K. Hansen, M. Ferita, H. M. Neuendorf, T. Liu, E. A. Smith, E. M. Mehrabad, M. Hejna, T. E. Moustafa, D. Lange, M. Hu, F. Vand-Rajabpour, A. Done, C. A. Becker, M. Lieberman, M. Chang, B. K. Lohman, C. J. Stubben, M. Q. Reeves, X. Zhang, L. S. Weinberger, M. W. VanBrocklin, D. C. Deacon, D. Grossman, B. T. Spike, A. Lex, G. M. Boyle, R. Kulkarni, T. A. Zangle, R. L. Judson-Torres.
A BRN2:MYC transcriptional axis regulates interconversion between therapy-resistant and tumorigenic phenotypes in melanoma, In Cell Reports, Vol. 44, No. 12, 2025.
Metastatic spread and therapeutic resistance are the principal causes of cancer mortality. For melanoma, these processes rely on the capacity of cells to switch between transcriptional states. Although targeting transcriptional states pharmacologically is promising, the mechanisms by which melanoma cells switch between states—and how these processes differ from melanocytes—remain poorly understood. Here, we isolate distinct melanoma states with unique phenotypes: a MYC-driven state, essential for tumor initiation yet sensitive to BRAF inhibition, and a dedifferentiated, invasive BRN2-high state enriched in therapy-resistant cells but not directly tumorigenic. Transitions between phenotypes occur through intermediate, more differentiated states. Unexpectedly, the BRN2-high state is also present in melanocytes, whereas the MYC state is exclusive to melanoma. Melanoma cells also exhibit an increased frequency of transitions across states. These findings highlight that accelerated phenotypic switching, rather than mere state diversity, is a defining feature of melanoma progression.
2024
J. Adams, K. Iyer, S. Elhabian.
Weakly Supervised Bayesian Shape Modeling from Unsegmented Medical Images, Subtitled arXiv:2405.09697v1, 2024.
Anatomical shape analysis plays a pivotal role in clinical research and hypothesis testing, where the relationship between form and function is paramount. Correspondence-based statistical shape modeling (SSM) facilitates population-level morphometrics but requires a cumbersome, potentially bias-inducing construction pipeline. Recent advancements in deep learning have streamlined this process in inference by providing SSM prediction directly from unsegmented medical images. However, the proposed approaches are fully supervised and require utilizing a traditional SSM construction pipeline to create training data, thus inheriting the associated burdens and limitations. To address these challenges, we introduce a weakly supervised deep learning approach to predict SSM from images using point cloud supervision. Specifically, we propose reducing the supervision associated with the state-of-the-art fully Bayesian variational information bottleneck DeepSSM (BVIB-DeepSSM) model. BVIB-DeepSSM is an effective, principled framework for predicting probabilistic anatomical shapes from images with quantification of both aleatoric and epistemic uncertainties. Whereas the original BVIB-DeepSSM method requires strong supervision in the form of ground truth correspondence points, the proposed approach utilizes weak supervision via point cloud surface representations, which are more readily obtainable. Furthermore, the proposed approach learns correspondence in a completely data-driven manner without prior assumptions about the expected variability in shape cohort. Our experiments demonstrate that this approach yields similar accuracy and uncertainty estimation to the fully supervised scenario while substantially enhancing the feasibility of model training for SSM construction.
J. Adams, S. Elhabian.
Point2SSM++: Self-Supervised Learning of Anatomical Shape Models from Point Clouds, Subtitled arXiv:2405.09707v1, 2024.
Correspondence-based statistical shape modeling (SSM) stands as a powerful technology for morphometric analysis in clinical research. SSM facilitates population-level characterization and quantification of anatomical shapes such as bones and organs, aiding in pathology and disease diagnostics and treatment planning. Despite its potential, SSM remains under-utilized in medical research due to the significant overhead associated with automatic construction methods, which demand complete, aligned shape surface representations. Additionally, optimization-based techniques rely on bias-inducing assumptions or templates and have prolonged inference times as the entire cohort is simultaneously optimized. To overcome these challenges, we introduce Point2SSM++, a principled, self-supervised deep learning approach that directly learns correspondence points from point cloud representations of anatomical shapes. Point2SSM++ is robust to misaligned and inconsistent input, providing SSM that accurately samples individual shape surfaces while effectively capturing population-level statistics. Additionally, we present principled extensions of Point2SSM++ to adapt it for dynamic spatiotemporal and multi-anatomy use cases, demonstrating the broad versatility of the Point2SSM++ framework. Furthermore, we present extensions of Point2SSM++ tailored for dynamic spatiotemporal and multi-anatomy scenarios, showcasing the broad versatility of the framework. Through extensive validation across diverse anatomies, evaluation metrics, and clinically relevant downstream tasks, we demonstrate Point2SSM++’s superiority over existing state-of-the-art deep learning models and traditional approaches. Point2SSM++ substantially enhances the feasibility of SSM generation and significantly broadens its array of potential clinical applications.
S.I. Adams-Tew, H. Odéen, D.L. Parker, C.C. Cheng, B. Madore, A. Payne, S. Joshi.
Physics Informed Neural Networks for Estimation of Tissue Properties from Multi-echo Configuration State MRI, In Medical Image Computing and Computer Assisted Intervention -- MICCAI 2024, Springer Nature Switzerland, pp. 502--511. 2024.
This work investigates the use of configuration state imaging together with deep neural networks to develop quantitative MRI techniques for deployment in an interventional setting. A physics modeling technique for inhomogeneous fields and heterogeneous tissues is presented and used to evaluate the theoretical capability of neural networks to estimate parameter maps from configuration state signal data. All tested normalization strategies achieved similar performance in estimating T2 and T2*. Varying network architecture and data normalization had substantial impacts on estimated flip angle and T1, highlighting their importance in developing neural networks to solve these inverse problems. The developed signal modeling technique provides an environment that will enable the development and evaluation of physics-informed machine learning techniques for MR parameter mapping and facilitate the development of quantitative MRI techniques to inform clinical decisions during MR-guided treatments.
T. M. Athawale, B. Triana, T. Kotha, D. Pugmire, P. Rosen.
A Comparative Study of the Perceptual Sensitivity of Topological Visualizations to Feature Variations, In IEEE Transactions on Visualization and Computer Graphics, Vol. 30, No. 1, pp. 1074-1084. Jan, 2024.
DOI: 10.1109/TVCG.2023.3326592
Color maps are a commonly used visualization technique in which data are mapped to optical properties, e.g., color or opacity. Color maps, however, do not explicitly convey structures (e.g., positions and scale of features) within data. Topology-based visualizations reveal and explicitly communicate structures underlying data. Although we have a good understanding of what types of features are captured by topological visualizations, our understanding of people’s perception of those features is not. This paper evaluates the sensitivity of topology-based isocontour, Reeb graph, and persistence diagram visualizations compared to a reference color map visualization for synthetically generated scalar fields on 2-manifold triangular meshes embedded in 3D. In particular, we built and ran a human-subject study that evaluated the perception of data features characterized by Gaussian signals and measured how effectively each visualization technique portrays variations of data features arising from the position and amplitude variation of a mixture of Gaussians. For positional feature variations, the results showed that only the Reeb graph visualization had high sensitivity. For amplitude feature variations, persistence diagrams and color maps demonstrated the highest sensitivity, whereas isocontours showed only weak sensitivity. These results take an important step toward understanding which topology-based tools are best for various data and task scenarios and their effectiveness in conveying topological variations as compared to conventional color mapping.
B. Aubert, N. Khan, F. Toupin, M. Pacheco, A. Morris.
Deformable Vertebra 3D/2D Registration from Biplanar X-Rays Using Particle-Based Shape Modelling, In Shape in Medical Imaging, Springer Nature Switzerland, pp. 33--47. 2024.
ISSN: 978-3-031-75291-9
Patient-specific 3D vertebra models are essential for accurately assessing the spinal deformities quantitatively in 3D and for surgical planning, including determining the optimal implant size and 3D positioning. Calibrated biplanar X-rays serve as an alternative to CT scans to generate the 3D models in a weight-bearing standing position. This paper presents an intensity-based 3D/2D registration method for vertebra statistical shape model (VSSM), incorporating two key elements: the particle-based shape modeling and an image domain transfer for efficient image matching. In the 3D/3D setting, the VSSMs reach a surface reconstruction error of less than 0.5 mm. For 3D reconstruction from biplanar X-rays, the root mean square point-to-surface are 1.05 mm for L1 to L4 vertebrae and 1.6 mm for the L5 vertebra. The particle-based VSSMs offer a significant balance between the model compactness and the reconstruction error, which is advantageous for deformable 3D/2D registration.
A.Z.B. Aziz, M.S.T. Karanam, T. Kataria, S.Y. Elhabian.
EfficientMorph: Parameter-Efficient Transformer-Based Architecture for 3D Image Registration, Subtitled arXiv preprint arXiv:2403.11026, 2024.
Transformers have emerged as the state-of-the-art architecture in medical image registration, outperforming convolutional neural networks (CNNs) by addressing their limited receptive fields and overcoming gradient instability in deeper models. Despite their success, transformer-based models require substantial resources for training, including data, memory, and computational power, which may restrict their applicability for end users with limited resources. In particular, existing transformer-based 3D image registration architectures face three critical gaps that challenge their efficiency and effectiveness. Firstly, while mitigating the quadratic complexity of full attention by focusing on local regions, window-based attention mechanisms often fail to adequately integrate local and global information. Secondly, feature similarities across attention heads that were recently found in multi-head attention architectures indicate a significant computational redundancy, suggesting that the capacity of the network could be better utilized to enhance performance. Lastly, the granularity of tokenization, a key factor in registration accuracy, presents a trade-off; smaller tokens improve detail capture at the cost of higher computational complexity, increased memory demands, and a risk of overfitting. Here, we propose EfficientMorph, a transformer-based architecture for unsupervised 3D image registration. It optimizes the balance between local and global attention through a plane-based attention mechanism, reduces computational redundancy via cascaded group attention, and captures fine details without compromising computational efficiency, thanks to a Hi-Res tokenization strategy complemented by merging operations. We compare the effectiveness of EfficientMorph on two public datasets, OASIS and IXI, against other state-of-the-art models. Notably, EfficientMorph sets a new benchmark for performance on the OASIS dataset with ∼16-27× fewer parameters.
J. Baker, Q. Wang, M. Berzins, T. Strohmer, B. Wang.
Monotone Operator Theory-Inspired Message Passing for Learning Long-Range Interaction on Graphs, In Proceedings of The 27th International Conference on Artificial Intelligence and Statistics, Vol. 238, Edited by Dasgupta, Sanjoy and Mandt, Stephan and Li, Yingzhen, PMLR, pp. 2233--2241. 2024.
Learning long-range interactions (LRI) between distant nodes is crucial for many graph learning tasks. Predominant graph neural networks (GNNs) rely on local message passing and struggle to learn LRI. In this paper, we propose DRGNN to learn LRI leveraging monotone operator theory. DRGNN contains two key components: (1) we use a full node similarity matrix beyond adjacency matrix – drawing inspiration from the personalized PageRank matrix – as the aggregation matrix for message passing, and (2) we implement message-passing on graphs using Douglas-Rachford splitting to circumvent prohibitive matrix inversion. We demonstrate that DRGNN surpasses various advanced GNNs, including Transformer-based models, on several benchmark LRI learning tasks arising from different application domains, highlighting its efficacy in learning LRI. Code is available at \urlhttps://github.com/Utah-Math-Data-Science/PR-inspired-aggregation.
Z. Bastiani, R.M. Kirby, J. Hochhalter, S. Zhe.
Complexity-Aware Deep Symbolic Regression with Robust Risk-Seeking Policy Gradients, Subtitled arXiv:2406.06751, 2024.
This paper proposes a novel deep symbolic regression approach to enhance the robustness and interpretability of data-driven mathematical expression discovery. Despite the success of the state-of-the-art method, DSR, it is built on recurrent neural networks, purely guided by data fitness, and potentially meet tail barriers, which can zero out the policy gradient and cause inefficient model updates. To overcome these limitations, we use transformers in conjunction with breadth-first-search to improve the learning performance. We use Bayesian information criterion (BIC) as the reward function to explicitly account for the expression complexity and optimize the trade-off between interpretability and data fitness. We propose a modified risk-seeking policy that not only ensures the unbiasness of the gradient, but also removes the tail barriers, thus ensuring effective updates from top performers. Through a series of benchmarks and systematic experiments, we demonstrate the advantages of our approach.
J.W. Beiriger, W. Tao, Z. Irgebay, J. Smetona, L. Dvoracek, N. Kass, A. Dixon, C. Zhang, M. Mehta, R. Whitaker, J. Goldstein.
A Longitudinal Analysis of Pre-and Post-Operative Dysmorphology in Metopic Craniosynostosis, In The Cleft Palate Craniofacial Journal, Sage, 2024.
DOI: 10.1177/10556656241237605
Objective
Design
Setting
Patients
Main Outcome Measures
Results
Conclusions
C.C. Berggren, D. Jiang, Y.F. Wang, J.A. Bergquist, L. Rupp, Z. Liu, R.S. MacLeod, A. Narayan, L. Timmins.
Influence of Material Parameter Variability on the Predicted Coronary Artery Biomechanical Environment via Uncertainty Quantification, Subtitled arXiv preprint arXiv:2401.15047, 2024.
Central to the clinical adoption of patient-specific modeling strategies is demonstrating that simulation results are reliable and safe. Indeed, simulation frameworks must be robust to uncertainty in model input(s), and levels of confidence should accompany results. In this study, we applied a coupled uncertainty quantification-finite element (FE) framework to understand the impact of uncertainty in vascular material properties on variability in predicted stresses. Univariate probability distributions were fit to material parameters derived from layer-specific mechanical behavior testing of human coronary tissue. Parameters were assumed to be probabilistically independent, allowing for efficient parameter ensemble sampling. In an idealized coronary artery geometry, a forward FE model for each parameter ensemble was created to predict tissue stresses under physiologic loading. An emulator was constructed within the UncertainSCI software using polynomial chaos techniques, and statistics and sensitivities were directly computed. Results demonstrated that material parameter uncertainty propagates to variability in predicted stresses across the vessel wall, with the largest dispersions in stress within the adventitial layer. Variability in stress was most sensitive to uncertainties in the anisotropic component of the strain energy function. Moreover, unary and binary interactions within the adventitial layer were the main contributors to stress variance, and the leading factor in stress variability was uncertainty in the stress-like material parameter that describes the contribution of the embedded fibers to the overall artery stiffness. Results from a patient-specific coronary model confirmed many of these findings. Collectively, these data highlight the impact of material property variation on uncertainty in predicted artery stresses and present a pipeline to explore and characterize forward model uncertainty in computational biomechanics.
J.A. Bergquist, B. Zenger, J. Brundage, R.S. MacLeod, T.J. Bunch, R. Shah, X. Ye, A. Lyons, M. Torre, R. Ranjan, T. Tasdizen, B.A. Steinberg.
Performance of Off-the-Shelf Machine Learning Architectures and Biases in Low Left Ventricular Ejection Fraction Detection, In Heart Rhythm O2, Vol. 5, No. 9, pp. 644 - 654. 2024.
Background
Objective
Methods
Results
Conclusions
J.A. Bergquist, D. Dade, B. Zenger, R.S. MacLeod, X. Ye, R. Ranjan, T. Tasdizen, B.A. Steinberg.
Machine Learning Prediction of Blood Potassium at Different Time Cutoffs, In Computing in Cardiology 2024, 2024.
Because serum potassium and ECG morphology changes exhibit a well-understood connection, and the timeline of ECG changes can be relatively quick, there is motivation to explore the sensitivity of ML based prediction of serum potassium using 12 lead ECG data with respect to the time between the ECG and potassium readings.
We trained a convolutional neural network to classify abnormal (serum potassium above 5 mEq/L) vs normal (serum potassium between 4 and 5 mEq/L) from the ECG alone. We compared training with ECGs and potassium measurements filtered to be within 1 hour, 30 minutes, and 15 minutes of each other. We explored scenarios that both leveraged all available data at each time cutoff as well as restricted data to match training set sizes across the time cutoffs. For each case, we trained five separate instances of our neural network to account for variability.
The 1 hour cutoff with all data resulted in an average area under the receiver operator curve (AUC) of 0.850 and a weighted accuracy of 76.3%, 15 minutes resulted in 0.814, 72.5%, and 30 minutes. Truncating the training sets to the same size as the 15 minute cutoff results in comparable average accuracy and AUC for all. Our future studies will continue to explore the performance of ML potassium predictions through investigations of failure cases, identification of biases, and explainability analyses.
K. Borkiewicz, E. Jensen, Y. Miao, S. Levy, J.P. Naiman, J. Carpenter, K.E. Isaacs.
Audience Reach of Scientific Data Visualizations in Planetarium-Screened Films, 2024.
Quantifying the global reach of planetarium dome shows presents significant challenges due to the lack of standardized viewership tracking mechanisms across diverse planetarium venues. We present an analysis of the global impact of dome shows, presenting data regarding four documentary films from a single visualization lab. Specifically, we designed and administered a viewership survey of four long-running shows that contained cinematic scientific visualizations. Reported survey data shows that between 1.2 - 2.6 million people have viewed these four films across the 68 responding planetariums (mean: 1.9 million). When we include estimates and extrapolate for the 315 planetariums that licensed these shows, we arrive at an estimate of 16.5 - 24.1 million people having seen these films (mean: 20.3 million).
O. Cankur, A. Tomar, D. Nichols, C. Scully-Allison, K. Isaacs, A. Bhatele.
Automated Programmatic Performance Analysis of Parallel Programs, Subtitled arXiv:2401.13150v1, 2024.
Developing efficient parallel applications is critical to advancing scientific development but requires significant performance analysis and optimization. Performance analysis tools help developers manage the increasing complexity and scale of performance data, but often rely on the user to manually explore low-level data and are rigid in how the data can be manipulated. We propose a Python-based API, Chopper, which provides high-level and flexible performance analysis for both single and multiple executions of parallel applications. Chopper facilitates performance analysis and reduces developer effort by providing configurable high-level methods for common performance analysis tasks such as calculating load imbalance, hot paths, scalability bottlenecks, correlation between metrics and CCT nodes, and causes of performance variability within a robust and mature Python environment that provides fluid access to lower-level data manipulations. We demonstrate how Chopper allows developers to quickly and succinctly explore performance and identify issues across applications such as AMG, Laghos, LULESH, Quicksilver and Tortuga.
Page 7 of 150
