SCI Publications
2024
M. Cooley, S. Zhe, R.M. Kirby, V. Shankar.
Polynomial-Augmented Neural Networks (PANNs) with Weak Orthogonality Constraints for Enhanced Function and PDE Approximation, Subtitled arXiv preprint arXiv:2406.02336, 2024.
We present polynomial-augmented neural networks (PANNs), a novel machine learning architecture that combines deep neural networks (DNNs) with a polynomial approximant. PANNs combine the strengths of DNNs (flexibility and efficiency in higher-dimensional approximation) with those of polynomial approximation (rapid convergence rates for smooth functions). To aid in both stable training and enhanced accuracy over a variety of problems, we present (1) a family of orthogonality constraints that impose mutual orthogonality between the polynomial and the DNN within a PANN; (2) a simple basis pruning approach to combat the curse of dimensionality introduced by the polynomial component; and (3) an adaptation of a polynomial preconditioning strategy to both DNNs and polynomials. We test the resulting architecture for its polynomial reproduction properties, ability to approximate both smooth functions and functions of limited smoothness, and as a method for the solution of partial differential equations (PDEs). Through these experiments, we demonstrate that PANNs offer superior approximation properties to DNNs for both regression and the numerical solution of PDEs, while also offering enhanced accuracy over both polynomial and DNN-based regression (each) when regressing functions with limited smoothness.
M. Cooley, R.M. Kirby, S. Zhe, V. Shankar.
HyResPINNs: Adaptive Hybrid Residual Networks for Learning Optimal Combinations of Neural and RBF Components for Physics-Informed Modeling, Subtitled arXiv:2410.03573, 2024.
Physics-informed neural networks (PINNs) are an increasingly popular class of techniques for the numerical solution of partial differential equations (PDEs), where neural networks are trained using loss functions regularized by relevant PDE terms to enforce physical constraints. We present a new class of PINNs called HyResPINNs, which augment traditional PINNs with adaptive hybrid residual blocks that combine the outputs of a standard neural network and a radial basis function (RBF) network. A key feature of our method is the inclusion of adaptive combination parameters within each residual block, which dynamically learn to weigh the contributions of the neural network and RBF network outputs. Additionally, adaptive connections between residual blocks allow for flexible information flow throughout the network. We show that HyResPINNs are more robust to training point locations and neural network architectures than traditional PINNs. Moreover, HyResPINNs offer orders of magnitude greater accuracy than competing methods on certain problems, with only modest increases in training costs. We demonstrate the strengths of our approach on challenging PDEs, including the Allen-Cahn equation and the Darcy-Flow equation. Our results suggest that HyResPINNs effectively bridge the gap between traditional numerical methods and modern machine learning-based solvers.
H. Csala, A. Mohan, D. Livescu, A. Arzani.
Physics-constrained coupled neural differential equations for one dimensional blood flow modeling, Subtitled arXiv:2411.05631, 2024.
Computational cardiovascular flow modeling plays a crucial role in understanding blood flow dynamics. While 3D models provide acute details, they are computationally expensive, especially with fluid-structure interaction (FSI) simulations. 1D models offer a computationally efficient alternative, by simplifying the 3D Navier-Stokes equations through axisymmetric flow assumption and cross-sectional averaging. However, traditional 1D models based on finite element methods (FEM) often lack accuracy compared to 3D averaged solutions. This study introduces a novel physics-constrained machine learning technique that enhances the accuracy of 1D cardiovascular flow models while maintaining computational efficiency. Our approach, utilizing a physics-constrained coupled neural differential equation (PCNDE) framework, demonstrates superior performance compared to conventional FEM-based 1D models across a wide range of inlet boundary condition waveforms and stenosis blockage ratios. A key innovation lies in the spatial formulation of the momentum conservation equation, departing from the traditional temporal approach and capitalizing on the inherent temporal periodicity of blood flow. This spatial neural differential equation formulation switches space and time and overcomes issues related to coupling stability and smoothness, while simplifying boundary condition implementation. The model accurately captures flow rate, area, and pressure variations for unseen waveforms and geometries. We evaluate the model’s robustness to input noise and explore the loss landscapes associated with the inclusion of different physics terms. This advanced 1D modeling technique offers promising potential for rapid cardiovascular simulations, achieving computational efficiency and accuracy. By combining the strengths of physics-based and data-driven modeling, this approach enables fast and accurate cardiovascular simulations.
H. Csala, O. Amili, R.M. D'Souza, A. Arzani.
A comparison of machine learning methods for recovering noisy and missing 4D flow MRI data, In Int J Numer Method Biomed Eng, Vol. 40, No. 11, 2024.
Experimental blood flow measurement techniques are invaluable for a better understanding of cardiovascular disease formation, progression, and treatment. One of the emerging methods is time-resolved three-dimensional phase-contrast magnetic resonance imaging (4D flow MRI), which enables noninvasive time-dependent velocity measurements within large vessels. However, several limitations hinder the usability of 4D flow MRI and other experimental methods for quantitative hemodynamics analysis. These mainly include measurement noise, corrupt or missing data, low spatiotemporal resolution, and other artifacts. Traditional filtering is routinely applied for denoising experimental blood flow data without any detailed discussion on why it is preferred over other methods. In this study, filtering is compared to different singular value decomposition (SVD)-based machine learning and autoencoder-type deep learning methods for denoising and filling in missing data (imputation). An artificially corrupted and voxelized computational fluid dynamics (CFD) simulation as well as in vitro 4D flow MRI data are used to test the methods. SVD-based algorithms achieve excellent results for the idealized case but severely struggle when applied to in vitro data. The autoencoders are shown to be versatile and applicable to all investigated cases. For denoising, the in vitro 4D flow MRI data, the denoising autoencoder (DAE), and the Noise2Noise (N2N) autoencoder produced better reconstructions than filtering both qualitatively and quantitatively. Deep learning methods such as N2N can result in noise-free velocity fields even though they did not use clean data during training. This work presents one of the first comprehensive assessments and comparisons of various classical and modern machine-learning methods for enhancing corrupt cardiovascular flow data in diseased arteries for both synthetic and experimental test cases.
D. Dade, J.A. Bergquist, R.S. MacLeod, X. Ye, R. Ranjan, B. Steinberg, T. Tasdizen.
A Survey of Augmentation Techniques for Enhancing ECG Representation Through Self-Supervised Contrastive Learning, In Computing in Cardiology 2024, 2024.
The electrocardiogram (ECG) is the most common non-invasive tool to measure the electrical activity of the heart and assess cardiac health. Despite their ubiquity and utility, traditional ECG analysis methods are limited in many impactful diseases. Machine learning tools can be employed to automate task-specific detection of diseases, and to detect patterns that are ignored by traditional ECG analysis. Contemporary machine learning tools are limited by requirements for large labeled datasets, which can be scarce for rare diseases. Self-supervised learning (SSL) can address this data scarcity. We implemented the momentum contrast(MoCo) framework, a form of SSL, using a large clinical ECG dataset. We then assessed the learning using Low Left Ventricular Ejection Fraction (LVEF) Detection as the downstream task. We compared the SSL improvement of LVEF classification across different input augmentations. We observed that optimal augmentation hyperparameters varied substantially based on the training dataset size, indicating that augmentation strategies may need to be tuned based on problem and dataset size.
H. Dai, S. Joshi.
Refining Skewed Perceptions in Vision-Language Models through Visual Representations, Subtitled arXiv preprint arXiv:2405.14030, 2024.
Large vision-language models (VLMs), such as CLIP, have become foundational, demonstrating remarkable success across a variety of downstream tasks. Despite their advantages, these models, akin to other foundational systems, inherit biases from the disproportionate distribution of real-world data, leading to misconceptions about the actual environment. Prevalent datasets like ImageNet are often riddled with non-causal, spurious correlations that can diminish VLM performance in scenarios where these contextual elements are absent. This study presents an investigation into how a simple linear probe can effectively distill task-specific core features from CLIP’s embedding for downstream applications. Our analysis reveals that the CLIP text representations are often tainted by spurious correlations, inherited in the biased pre-training dataset. Empirical evidence suggests that relying on visual representations from CLIP, as opposed to text embedding, is more practical to refine the skewed perceptions in VLMs, emphasizing the superior utility of visual representations in overcoming embedded biases
S. Dasetty, T.C. Bidone, A.L. Ferguson.
Data-driven prediction of αIIbβ3 integrin activation pathways using nonlinear manifold learning and deep generative modeling, In Biophysical Journal, Vol. 123, 2024.
The integrin heterodimer is a transmembrane protein critical for driving cellular process and is a therapeutic target in the treatment of multiple diseases linked to its malfunction. Activation of integrin involves conformational transitions between bent and extended states. Some of the conformations that are intermediate between bent and extended states of the heterodimer have been experimentally characterized, but the full activation pathways remain unresolved both experimentally due to their transient nature and computationally due to the challenges in simulating rare barrier crossing events in these large molecular systems. An understanding of the activation pathways can provide new fundamental understanding of the biophysical processes associated with the dynamic interconversions between bent and extended states and can unveil new putative therapeutic targets. In this work, we apply nonlinear manifold learning to coarse-grained molecular dynamics simulations of bent, extended, and two intermediate states of aIIbb3 integrin to learn a low-dimensional embedding of the configurational phase space. We then train deep generative models to learn an inverse mapping between the low-dimensional embedding and high-dimensional molecular space and use these models to interpolate the molecular configurations constituting the activation pathways between the experimentally characterized states. This work furnishes plausible predictions of integrin activation pathways and reports a generic and transferable multiscale technique to predict transition pathways for biomolecular systems.
Y.S. Dogrusoz, L. Bear, J.A. Bergquist, A. Rababah, W. Good, J. Stoks, J. Svehikova, E. van Dam, D.H. Brooks, R.S. MacLeod.
Evaluation of five methods for the interpolation of bad leads in the solution of the inverse electrocardiography problem, In Physiological Measurement, Vol. 24, No. 45, 2024.
DOI: 10.1088/1361-6579/ad74d6
Objective.This study aims to assess the sensitivity of epicardial potential-based electrocardiographic imaging (ECGI) to the removal or interpolation of bad leads.Approach.We utilized experimental data from two distinct centers. Langendorff-perfused pig (n= 2) and dog (n= 2) hearts were suspended in a human torso-shaped tank and paced from the ventricles. Six different bad lead configurations were designed based on clinical experience. Five interpolation methods were applied to estimate the missing data. Zero-order Tikhonov regularization was used to solve the inverse problem for complete data, data with removed bad leads, and interpolated data. We assessed the quality of interpolated ECG signals and ECGI reconstructions using several metrics, comparing the performance of interpolation methods and the impact of bad lead removal versus interpolation on ECGI.
Main results.The performance of ECG interpolation strongly correlated with ECGI reconstruction. The hybrid method exhibited the best performance among interpolation techniques, followed closely by the inverse-forward and Kriging methods. Bad leads located over high amplitude/high gradient areas on the torso significantly impacted ECGI reconstructions, even with minor interpolation errors. The choice between removing or interpolating bad leads depends on the location of missing leads and confidence in interpolation performance. If uncertainty exists, removing bad leads is the safer option, particularly when they are positioned in high amplitude/high gradient regions. In instances where interpolation is necessary, the inverse-forward and Kriging methods, which do not require training, are recommended.
Significance.This study represents the first comprehensive evaluation of the advantages and drawbacks of interpolating versus removing bad leads in the context of ECGI, providing valuable insights into ECGI performance.
J. Dong, E. Kwan, J.A. Bergquist, B.A. Steinberg, D.J. Dosdall, E. DiBella, R.S. MacLeod, T.J. Bunch, R. Ranjan.
Ablation-induced left atrial mechanical dysfunction recovers in weeks after ablation, In Journal of Interventional Cardiac Electrophysiology, Springer, 2024.
Background
The immediate impact of catheter ablation on left atrial mechanical function and the timeline for its recovery in patients undergoing ablation for atrial fibrillation (AF) remain uncertain. The mechanical function response to catheter ablation in patients with different AF types is poorly understood.
Methods
A total of 113 AF patients were included in this retrospective study. Each patient had three magnetic resonance imaging (MRI) studies in sinus rhythm: one pre-ablation, one immediate post-ablation (within 2 days after ablation), and one post-ablation follow-up MRI (≤ 3 months). We used feature tracking in the MRI cine images to determine peak longitudinal atrial strain (PLAS). We evaluated the change in strain from pre-ablation, immediately after ablation to post-ablation follow-up in a short-term study (< 50 days) and a 3-month study (3 months after ablation).
Results
The PLAS exhibited a notable reduction immediately after ablation, compared to both pre-ablation levels and those observed in follow-up studies conducted at short-term (11.1 ± 9.0 days) and 3-month (69.6 ± 39.6 days) intervals. However, there was no difference between follow-up and pre-ablation PLAS. The PLAS returned to 95% pre-ablation level within 10 days. Paroxysmal AF patients had significantly higher pre-ablation PLAS than persistent AF patients in pre-ablation MRIs. Both type AF patients had significantly lower immediate post-ablation PLAS compared with pre-ablation and post-ablation PLAS.
Conclusion
The present study suggested a significant drop in PLAS immediately after ablation. Left atrial mechanical function recovered within 10 days after ablation. The drop in PLAS did not show a substantial difference between paroxysmal and persistent AF patients.
J. Dong, E. Kwan, J.A. Bergquist, D.J. Dosdall, E.V. DiBella, R.S. MacLeod, G. Stoddard, K. Konstantidinis, B.A. Steinberg, T.J. Bunch, R. Ranjan.
Left atrial functional changes associated with repeated catheter ablations for atrial fibrillation, In J Cardiovasc Electrophysiol, 2024.
DOI: 10.1111/jce.16484
PubMed ID: 39474660
Introduction: The impact of repeated atrial fibrillation (AF) ablations on left atrial (LA) mechanical function remains uncertain, with limited long-term follow-up data.
Methods: This retrospective study involved 108 AF patients who underwent two catheter ablations with cardiac magnetic resonance imaging (MRI) done before and 3 months after each of the ablations from 2010 to 2021. The rate of change in peak longitudinal atrial strain (PLAS) assessed LA function. Additionally, a sub-study of 36 patients who underwent an extra MRI before the second ablation, gave us an additional time segment to evaluate the basis of change in PLAS.
Results: In the two-ablation, three MRI sub-study 1, the PLAS percent change rate was similar before and after the first ablation (r11 = -0.9 ± 3.1%/year, p = 0.771). However, the strain change rate from postablation 1 to postablation 2 was significantly worse (r12 = -23.7 ± 4.8%/year, p < 0.001). In the sub-study 2 with four MRIs, all three rates were negative, with reductions from postablation 1 to pre-ablation 2 (r22 = -13.3 ± 2.6%/year, p < 0.001) and from pre-ablation 2 to postablation 2 (r23 = -8.9 ± 3.9%/year, p = 0.028) being significant.
Conclusion: The present study suggests that the more ablations performed, the more significant the decrease in the postablation mechanical function of the LA. The natural progression of AF (strain change from postablation 1 to pre-ablation 2) had a greater negative influence on LA mechanical function compared to the second ablation itself suggesting that second ablation in patients with recurrence after first ablation is an effective strategy even from the LA mechanical function aspect.
S. Dubey, Y. Chong, B. Knudsen, S.Y. Elhabian.
VIMs: Virtual Immunohistochemistry Multiplex staining via Text-to-Stain Diffusion Trained on Uniplex Stains, Subtitled arXiv:2407.19113, 2024.
This paper introduces a Virtual Immunohistochemistry Multiplex staining (VIMs) model designed to generate multiple immunohistochemistry (IHC) stains from a single hematoxylin and eosin (H&E) stained tissue section. IHC stains are crucial in pathology practice for resolving complex diagnostic questions and guiding patient treatment decisions. While commercial laboratories offer a wide array of up to 400 different antibody-based IHC stains, small biopsies often lack sufficient tissue for multiple stains while preserving material for subsequent molecular testing. This highlights the need for virtual IHC staining. Notably, VIMs is the first model to address this need, leveraging a large vision-language single-step diffusion model for virtual IHC multiplexing through text prompts for each IHC marker. VIMs is trained on uniplex paired H&E and IHC images, employing an adversarial training module. Testing of VIMs includes both paired and unpaired image sets. To enhance computational efficiency, VIMs utilizes a pre-trained large latent diffusion model fine-tuned with small, trainable weights through the Low-Rank Adapter (LoRA) approach. Experiments on nuclear and cytoplasmic IHC markers demonstrate that VIMs outperforms the base diffusion model and achieves performance comparable to Pix2Pix, a standard generative model for paired image translation. Multiple evaluation methods, including assessments by two pathologists, are used to determine the performance of VIMs. Additionally, experiments with different prompts highlight the impact of text conditioning. This paper represents the first attempt to accelerate histopathology research by demonstrating the generation of multiple IHC stains from a single H&E input using a single model trained solely on uniplex data. This approach relaxes the traditional need for multiplex training sets, significantly broadening the applicability and accessibility of virtual IHC staining techniques.
K. Eckelt, K. Gadhave, A. Lex, M. Streit.
Loops: Leveraging Provenance and Visualization to Support Exploratory Data Analysis in Notebooks, In Computer Graphics Forum, 2024.
Exploratory data science is an iterative process of obtaining, cleaning, profiling, analyzing, and interpreting data. This cyclical way of working creates challenges within the linear structure of computational notebooks, leading to issues with code quality, recall, and reproducibility. To remedy this, we present Loops, a set of visual support techniques for iterative and exploratory data analysis in computational notebooks. Loops leverages provenance information to visualize the impact of changes made within a notebook. In visualizations of the notebook provenance, we trace the evolution of the notebook over time and highlight differences between versions. Loops visualizes the provenance of code, markdown, tables, visualizations, and images and their respective differences. Analysts can explore these differences in detail in a separate view. Loops not only improves the reproducibility of notebooks but also supports analysts in their data science work by showing the effects of changes and facilitating comparison of multiple versions. We demonstrate our approach’s utility and potential impact in two use cases and feedback from notebook users from various backgrounds.
G. Eisenhauer, N. Podhorszki, A. Gainaru, S. Klasky, M. Parashar, M. Wolf, E. Suchtya, E. Fredj, V. Bolea, F. Poschel, K. Steiniger, M. Bussmann, R. Pausch, S. Chandrasekaran.
Streaming Data in HPC Workflows Using ADIOS, Subtitled arXiv:2410.00178v1, 2024.
The “IO Wall” problem, in which the gap between computation rate and data access rate grows continuously, poses significant problems to scientific workflows which have traditionally relied upon using the filesystem for intermediate storage between workflow stages. One way to avoid this problem in scientific workflows is to stream data directly from producers to consumers and avoiding storage entirely. However, the manner in which this is accomplished is key to both performance and usability. This paper presents the Sustainable Staging Transport, an approach which allows direct streaming between traditional file writers and readers with few application changes. SST is an ADIOS “engine”, accessible via standard ADIOS APIs, and because ADIOS allows engines to be chosen at run-time, many existing file-oriented ADIOS workflows can utilize SST for direct application-to-application communication without any source code changes. This paper describes the design of SST and presents performance results from various applications that use SST, for feeding model training with simulation data with substantially higher bandwidth than the theoretical limits of Frontier’s file system, for strong coupling of separately developed applications for multiphysics multiscale simulation, or for in situ analysis and visualization of data to complete all data processing shortly after the simulation finishes.
I.J. Eliza, J. Wagoner, J. Wilburn, N. Lanza, D. Hajas, A. Lex.
Accessible Text Descriptions for UpSet Plots, In 2024 1st Workshop on Accessible Data Visualization (AccessViz), pp. 1--4. 2024.
Data visualizations are typically not accessible to blind and low-vision users. The most widely used remedy for making data visualizations accessible is text descriptions. Yet, manually creating useful text descriptions is often omitted by visualization authors, either because of a lack of awareness or a perceived burden. Automatically generated text descriptions are a potential partial remedy. However, with current methods it is unfeasible to create text descriptions for complex scientific charts. In this paper, we describe our methods for generating text descriptions for one complex scientific visualization: the UpSet plot. UpSet is a widely used technique for the visualization and analysis of sets and their intersections. At the same time, UpSet is arguably unfamiliar to novices and used mostly in scientific contexts. Generating text descriptions for UpSet plots is challenging because the patterns observed in UpSet plots have not been studied. We first analyze patterns present in dozens of published UpSet plots. We then introduce software that generates text descriptions for UpSet plots based on the patterns present in the chart. Finally, we introduce a web service that generates text descriptions based on a specification of an UpSet plot, and demonstrate its use in both an interactive web-based implementation and a static Python implementation of UpSet.
Y. Epshteyn, A. Narayan, Y. Yu.
Energy Stable and Structure-Preserving Algorithms for the Stochastic Galerkin System of 2D Shallow Water Equations, Subtitled arXiv:2412.16353, 2024.
Shallow water equations (SWE) are fundamental nonlinear hyperbolic PDE-based models in fluid dynamics that are essential for studying a wide range of geophysical and engineering phenomena. Therefore, stable and accurate numerical methods for SWE are needed. Although some algorithms are well studied for deterministic SWE, more effort should be devoted to handling the SWE with uncertainty. In this paper, we incorporate uncertainty through a stochastic Galerkin (SG) framework, and building on an existing hyperbolicity-preserving SG formulation for 2D SWE, we construct the corresponding entropy flux pair, and develop structure-preserving, well-balanced, second-order energy conservative and energy stable finite volume schemes for the SG formulation of the two-dimensional shallow water system. We demonstrate the efficacy, applicability, and robustness of these structure-preserving algorithms through several challenging numerical experiments.
A. Ferrero, E. Ghelichkhan, H. Manoochehri, M.M. Ho, D.J. Albertson, B.J. Brintz, T. Tasdizen, R.T. Whitaker, B. Knudsen.
HistoEM: A Pathologist-Guided and Explainable Workflow Using Histogram Embedding for Gland Classification, In Modern Pathology, Vol. 37, No. 4, 2024.
Pathologists have, over several decades, developed criteria for diagnosing and grading prostate cancer. However, this knowledge has not, so far, been included in the design of convolutional neural networks (CNN) for prostate cancer detection and grading. Further, it is not known whether the features learned by machine-learning algorithms coincide with diagnostic features used by pathologists. We propose a framework that enforces algorithms to learn the cellular and subcellular differences between benign and cancerous prostate glands in digital slides from hematoxylin and eosin–stained tissue sections. After accurate gland segmentation and exclusion of the stroma, the central component of the pipeline, named HistoEM, utilizes a histogram embedding of features from the latent space of the CNN encoder. Each gland is represented by 128 feature-wise histograms that provide the input into a second network for benign vs cancer classification of the whole gland. Cancer glands are further processed by a U-Net structured network to separate low-grade from high-grade cancer. Our model demonstrates similar performance compared with other state-of-the-art prostate cancer grading models with gland-level resolution. To understand the features learned by HistoEM, we first rank features based on the distance between benign and cancer histograms and visualize the tissue origins of the 2 most important features. A heatmap of pixel activation by each feature is generated using Grad-CAM and overlaid on nuclear segmentation outlines. We conclude that HistoEM, similar to pathologists, uses nuclear features for the detection of prostate cancer. Altogether, this novel approach can be broadly deployed to visualize computer-learned features in histopathology images.
S. Garg, J. Zhang, R. Pitchumani, M. Parashar, B. Xie, S. Kannan.
CrossPrefetch: Accelerating I/O Prefetching for Modern Storage, In 29th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, ACM, 2024.
We introduce CrossPrefetch, a novel cross-layered I/O prefetching mechanism that operates across the OS and a user-level runtime to achieve optimal performance. Existing OS prefetching mechanisms suffer from rigid interfaces that do not provide information to applications on the prefetch effectiveness, suffer from high concurrency bottlenecks, and are inefficient in utilizing available system memory. CrossPrefetch addresses these limitations by dividing responsibilities between the OS and runtime, minimizing overhead, and achieving low cache misses, lock contentions, and higher I/O performance.
CrossPrefetch tackles the limitations of rigid OS prefetching interfaces by maintaining and exporting cache state and prefetch effectiveness to user-level runtimes. It also addresses scalability and concurrency bottlenecks by distinguishing between regular I/O and prefetch operations paths and introduces fine-grained prefetch indexing for shared files. Finally, CrossPrefetch designs low-interference access pattern prediction combined with support for adaptive and aggressive techniques to exploit memory capacity and storage bandwidth. Our evaluation of CrossPrefetch, encompassing microbenchmarks, macrobenchmarks, and real-world workloads, illustrates performance gains of up to 1.22x- 3.7x in I/O throughput. We also evaluate CrossPrefetch across different file systems and local and remote storage configurations.
E. Gasparovic, E. Purvine, R. Sazdanovic, B. Wang, Y. Wang, L. Ziegelmeier.
A survey of simplicial, relative, and chain complex homology theories for hypergraphs, 2024.
Hypergraphs have seen widespread applications in network and data science communities in recent years. We present a survey of recent work to define topological objects from hypergraphs—specifically simplicial, relative, and chain complexes—that can be used to build homology theories for hypergraphs. We define and describe nine different homology theories and their relevant topological objects. We discuss some interesting properties of each method to show how the hypergraph structures are preserved or destroyed by modifying a hypergraph. Finally, we provide a series of illustrative examples by computing many of these homology theories for small hypergraphs to show the variability of the methods and build intuition.
S. Ghorbany, M. Hu, S. Yao, C. Wang, Q.C. Nguyen, X. Yue, M. Alirezaei, T. Tasdizen, M Sisk.
Examining the role of passive design indicators in energy burden reduction: Insights from a machine learning and deep learning approach, In Building and Environment, Elsevier, 2024.
Passive design characteristics (PDC) play a pivotal role in reducing the energy burden on households without imposing additional financial constraints on project stakeholders. However, the scarcity of PDC data has posed a challenge in previous studies when assessing their energy-saving impact. To tackle this issue, this research introduces an innovative approach that combines deep learning-powered computer vision with machine learning techniques to examine the relationship between PDC and energy burden in residential buildings. In this study, we employ a convolutional neural network computer vision model to identify and measure key indicators, including window-to-wall ratio (WWR), external shading, and operable window types, using Google Street View images within the Chicago metropolitan area as our case study. Subsequently, we utilize the derived passive design features in conjunction with demographic characteristics to train and compare various machine learning methods. These methods encompass Decision Tree Regression, Random Forest Regression, and Support Vector Regression, culminating in the development of a comprehensive model for energy burden prediction. Our framework achieves a 74.2 % accuracy in forecasting the average energy burden. These results yield invaluable insights for policymakers and urban planners, paving the way toward the realization of smart and sustainable cities.
C. Gritton.
MODELING ELECTROCHEMISTRY PROBLEMS USING THE MATERIAL POINT AND FINITE VOLUME METHODS, Subtitled Ph.D. Dissertation, School of Computing, 2024.
The demand for batteries is continually growing. The use of personal electronic devices such as smart phones and tablets, the increase in renewable energy sources, and the continuing trend to replace fossil-fuel-based personal transport with electric vehicles all require the use of batteries. With this increase in demand is the need for lighter and more efficient batteries. Computational modeling plays a key role in the development of new designs and the exploration of novel of materials, which in turn leads to improvements in battery efficiency and weight. The modeling of a battery cell involves capturing the physics of electrochemical diffusion, deformation, and electrostatics. The multiphysics nature, as well as the difference in time scales at which the physical processes occur, make this problem difficult. The numerical method used to model the multiphysics of an battery cell undergoing a charge/disscharge cycle is the Material Point Method (MPM). This work focuses on two key areas: improving upon current approaches to MPM and applying MPM and the finite volume method to battery modeling. The improvements in MPM have been achieved in the following ways: 1) developing approaches by which information is transferred back and forth between particles and the grid through the use of nullspace filters and corrected derivatives; 2) analyzing different time integration schemes that can be used to address the multiscale nature of modeling the coupled diffusion/deformation problem using MPM; and 3) the use MPM in the reference configuration to solve diffusion type problems. The application of MPM to modeling battery problems has been demonstrated through the simulation of a silicon electrode within a lithium ion battery undergoing multiple charge/discharge cycles. The finite volume vethod is applied to modeling a full battery cell during the charging phase, which involves the coupled physics of electrochemical diffusion and electrostatics.
Page 8 of 149
