SCI Publications
2016
E. Wong, S. Palande, Bei Wang, B. Zielinski, J. Anderson, P. T. Fletcher.
Kernel Partial Least Squares Regression for Relating Functional Brain Network Topology to Clinical Measures of Behavior, In 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI), IEEE, April, 2016.
DOI: 10.1109/isbi.2016.7493506
In this paper we present a novel method for analyzing the relationship between functional brain networks and behavioral phenotypes. Drawing from topological data analysis, we first extract topological features using persistent homology from functional brain networks that are derived from correlations in resting-state fMRI. Rather than fixing a discrete network topology by thresholding the connectivity matrix, these topological features capture the network organization across all continuous threshold values. We then propose to use a kernel partial least squares (kPLS) regression to statistically quantify the relationship between these topological features and behavior measures. The kPLS also provides an elegant way to combine multiple image features by using linear combinations of multiple kernels. In our experiments we test the ability of our proposed brain network analysis to predict autism severity from rs-fMRI. We show that combining correlations with topological features gives better prediction of autism severity than using correlations alone.
2015
K.K. Aras, W. Good, J. Tate, B.M. Burton, D.H. Brooks, J. Coll-Font, O. Doessel, W. Schulze, D. Patyogaylo, L. Wang, P. Van Dam,, R.S. MacLeod.
Experimental Data and Geometric Analysis Repository: EDGAR, In Journal of Electrocardiology, 2015.
Introduction
The "Experimental Data and Geometric Analysis Repository", or EDGAR is an Internet-based archive of curated data that are freely distributed to the international research community for the application and validation of electrocardiographic imaging (ECGI) techniques. The EDGAR project is a collaborative effort by the Consortium for ECG Imaging (CEI, ecg-imaging.org), and focused on two specific aims. One aim is to host an online repository that provides access to a wide spectrum of data, and the second aim is to provide a standard information format for the exchange of these diverse datasets.
Methods
The EDGAR system is composed of two interrelated components: 1) a metadata model, which includes a set of descriptive parameters and information, time signals from both the cardiac source and body-surface, and extensive geometric information, including images, geometric models, and measure locations used during the data acquisition/generation; and 2) a web interface. This web interface provides efficient, search, browsing, and retrieval of data from the repository.
Results
An aggregation of experimental, clinical and simulation data from various centers is being made available through the EDGAR project including experimental data from animal studies provided by the University of Utah (USA), clinical data from multiple human subjects provided by the Charles University Hospital (Czech Republic), and computer simulation data provided by the Karlsruhe Institute of Technology (Germany).
Conclusions
It is our hope that EDGAR will serve as a communal forum for sharing and distribution of cardiac electrophysiology data and geometric models for use in ECGI research.
I. Arganda-Carreras, S.C. Turaga, D.R. Berger, D. Ciresan, A. Giusti, L.M. Gambardella, J. Schmidhuber, D. Laptev, S. Dwivedi, J. Buhmann, T. Liu, M. Seyedhosseini, T. Tasdizen, L. Kamentsky, R. Burget, V. Uher, X. Tan, C. Sun, T.D. Pham, E. Bas, M.G. Uzunbas, A. Cardona, J. Schindelin, H.S. Seung.
Crowdsourcing the creation of image segmentation algorithms for connectomics, In Frontiers in Neuroanatomy, Vol. 9, Frontiers Media SA, Nov, 2015.
DOI: 10.3389/fnana.2015.00142
To stimulate progress in automating the reconstruction of neural circuits, we organized the first international challenge on 2D segmentation of electron microscopic (EM) images of the brain. Participants submitted boundary maps predicted for a test set of images, and were scored based on their agreement with a consensus of human expert annotations. The winning team had no prior experience with EM images, and employed a convolutional network. This "deep learning" approach has since become accepted as a standard for segmentation of EM images. The challenge has continued to accept submissions, and the best so far has resulted from cooperation between two teams. The challenge has probably saturated, as algorithms cannot progress beyond limits set by ambiguities inherent in 2D scoring and the size of the test dataset. Retrospective evaluation of the challenge scoring system reveals that it was not sufficiently robust to variations in the widths of neurite borders. We propose a solution to this problem, which should be useful for a future 3D segmentation challenge.
J. Bennett, F. Vivodtzev, V. Pascucci (Eds.).
Topological and Statistical Methods for Complex Data, Subtitled Tackling Large-Scale, High-Dimensional, and Multivariate Data Spaces, Mathematics and Visualization, Springer Berlin Heidelberg, 2015.
ISBN: 978-3-662-44899-1
This book contains papers presented at the Workshop on the Analysis of Large-scale,
High-Dimensional, and Multi-Variate Data Using Topology and Statistics, held in Le Barp,
France, June 2013. It features the work of some of the most prominent and recognized
leaders in the field who examine challenges as well as detail solutions to the analysis of
extreme scale data.
The book presents new methods that leverage the mutual strengths of both topological
and statistical techniques to support the management, analysis, and visualization
of complex data. It covers both theory and application and provides readers with an
overview of important key concepts and the latest research trends.
Coverage in the book includes multi-variate and/or high-dimensional analysis techniques,
feature-based statistical methods, combinatorial algorithms, scalable statistics algorithms,
scalar and vector field topology, and multi-scale representations. In addition, the book
details algorithms that are broadly applicable and can be used by application scientists to
glean insight from a wide range of complex data sets.
J. Bennett, R. Clay, G. Baker, M. Gamell, D. Hollman, S. Knight, H. Kolla, G. Sjaardema, N. Slattengren, K. Teranishi, J. Wilke, M. Bettencourt, S. Bova, K. Franko, P. Lin, R. Grant, S. Hammond, S. Olivier.
ASC ATDM Level 2 Milestone #5325, Subtitled Asynchronous Many-Task Runtime System Analysis and Assessment for Next Generation Platforms, Note: Sandia Report, 2015.
This report provides in-depth information and analysis to help create a technical road map for developing nextgeneration programming models and runtime systems that support Advanced Simulation and Computing (ASC) workload requirements. The focus herein is on asynchronous many-task (AMT) model and runtime systems, which are of great interest in the context of "exascale" computing, as they hold the promise to address key issues associated with future extreme-scale computer architectures. This report includes a thorough qualitative and quantitative examination of three best-of-class AMT runtime systems—Charm++, Legion, and Uintah, all of which are in use as part of the ASC Predictive Science Academic Alliance Program II (PSAAP-II) Centers. The studies focus on each of the runtimes' programmability, performance, and mutability. Through the experiments and analysis presented, several overarching findings emerge. From a performance perspective, AMT runtimes show tremendous potential for addressing extremescale challenges. Empirical studies show an AMT runtime can mitigate performance heterogeneity inherent to the machine itself and that Message Passing Interface (MPI) and AMT runtimes perform comparably under balanced conditions. From a programmability and mutability perspective however, none of the runtimes in this study are currently ready for use in developing production-ready Sandia ASC applications. The report concludes by recommending a codesign path forward, wherein application, programming model, and runtime system developers work together to define requirements and solutions. Such a requirements-driven co-design approach benefits the high-performance computing (HPC) community as a whole, with widespread community engagement mitigating risk for both application developers and runtime system developers.
J. Bennett, R. Clay, G. Baker, M. Gamell, D. Hollman, S. Knight, H. Kolla, G. Sjaardema, N. Slattengren, K. Teranishi, J. Wilke, M. Bettencourt, S. Bova, K. Franko, P. Lin, R. Grant, S. Hammond, S. Olivier, L. Kale, N. Jain, E. Mikida, A. Aiken, M. Bauer, W. Lee, E. Slaughter, S. Treichler, M. Berzins, T. Harman, A. Humphrey, J. Schmidt, D. Sunderland, P. McCormick, S. Gutierrez, M. Schulz, A. Bhatele, D. Boehme, P. Bremer, T. Gamblin.
ASC ATDM level 2 milestone #5325: Asynchronous many-task runtime system analysis and assessment for next generation platforms, Sandia National Laboratories, 2015.
This report provides in-depth information and analysis to help create a technical road map for developing nextgeneration programming models and runtime systems that support Advanced Simulation and Computing (ASC) workload requirements. The focus herein is on asynchronous many-task (AMT) model and runtime systems, which are of great interest in the context of "exascale" computing, as they hold the promise to address key issues associated with future extreme-scale computer architectures. This report includes a thorough qualitative and quantitative examination of three best-of-class AMT runtime systems—Charm++, Legion, and Uintah, all of which are in use as part of the ASC Predictive Science Academic Alliance Program II (PSAAP-II) Centers. The studies focus on each of the runtimes' programmability, performance, and mutability. Through the experiments and analysis presented, several overarching findings emerge. From a performance perspective, AMT runtimes show tremendous potential for addressing extremescale challenges. Empirical studies show an AMT runtime can mitigate performance heterogeneity inherent to the machine itself and that Message Passing Interface (MPI) and AMT runtimes perform comparably under balanced conditions. From a programmability and mutability perspective however, none of the runtimes in this study are currently ready for use in developing production-ready Sandia ASC applications. The report concludes by recommending a codesign path forward, wherein application, programming model, and runtime system developers work together to define requirements and solutions. Such a requirements-driven co-design approach benefits the high-performance computing (HPC) community as a whole, with widespread community engagement mitigating risk for both application developers and runtime system developers.
H. Bhatia, Bei Wang, G. Norgard, V. Pascucci, P. T. Bremer.
Local, Smooth, and Consistent Jacobi Set Simplification, In Computational Geometry, Vol. 48, No. 4, Elsevier, pp. 311-332. May, 2015.
DOI: 10.1016/j.comgeo.2014.10.009
This paper introduces the theoretical foundations of a new simplification framework for Jacobi sets. We present a new interpretation of Jacobi set simplification based on the perspective of domain segmentation. Generalizing the cancellation of critical points from scalar functions to Jacobi sets, we focus on simplifications that can be realized by smooth approximations of the corresponding functions, and show how these cancellations imply simultaneous simplification of contiguous subsets of the Jacobi set. Using these extended cancellations as atomic operations, we introduce an algorithm to successively cancel subsets of the Jacobi set with minimal modifications to some userdefined metric. We show that for simply connected domains, our algorithm reduces a given Jacobi set to its minimal configuration, that is, one with no birth-death points (a birth-death point is a specific type of singularity within the Jacobi set where the level sets of the two functions and the Jacobi set have a common normal direction).
P. T. Bremer, D. Maljovec, A. Saha, Bei Wang, J. Gaffney, B. K. Spears, V. Pascucci.
ND2AV: N-Dimensional Data Analysis and Visualization -- Analysis for the National Ignition Campaign, In Computing and Visualization in Science, 2015.
T. Bregman, R. Reznikov, M. Diwan, R. Raymond, C. R.Butson, J. N.Nobrega, C. Hamani.
Antidepressant-like Effects of Medial Forebrain Bundle Deep Brain Stimulation in Rats are not Associated With Accumbens Dopamine Release, In Brain Stimulation, Vol. 8, No. 4, pp. 708--713. 2015.
BACKGROUND:
Medial forebrain bundle (MFB) deep brain stimulation (DBS) is currently being investigated in patients with treatment-resistant depression. Striking features of this therapy are the large number of patients who respond to treatment and the rapid nature of the antidepressant response.
OBJECTIVE:
To study antidepressant-like behavioral responses, changes in regional brain activity, and monoamine release in rats receiving MFB DBS.
METHODS:
Antidepressant-like effects of MFB stimulation at 100 μA, 90 μs and either 130 Hz or 20 Hz were characterized in the forced swim test (FST). Changes in the expression of the immediate early gene (IEG) zif268 were measured with in situ hybridization and used as an index of regional brain activity. Microdialysis was used to measure DBS-induced dopamine and serotonin release in the nucleus accumbens.
RESULTS:
Stimulation at parameters that approximated those used in clinical practice, but not at lower frequencies, induced a significant antidepressant-like response in the FST. In animals receiving MFB DBS at high frequency, increases in zif268 expression were observed in the piriform cortex, prelimbic cortex, nucleus accumbens shell, anterior regions of the caudate/putamen and the ventral tegmental area. These structures are involved in the neurocircuitry of reward and are also connected to other brain areas via the MFB. At settings used during behavioral tests, stimulation did not induce either dopamine or serotonin release in the nucleus accumbens.
CONCLUSIONS:
These results suggest that MFB DBS induces an antidepressant-like effect in rats and recruits structures involved in the neurocircuitry of reward without affecting dopamine release in the nucleus accumbens.
C. R. Butson, C. C. McIntyre.
The use of stimulation field models for deep brain stimulation programming, In Brain Stimulation, Vol. 8, No. 5, Elsevier BV, pp. 976--978. September, 2015.
DOI: 10.1016/j.brs.2015.06.005
H. Carr, Z. Geng, J. Tierny, A. Chattophadhyay,, A. Knoll.
Fiber Surfaces: Generalizing Isosurfaces to Bivariate Data, In Computer Graphics Forum, Vol. 34, No. 3, pp. 241-250. 2015.
CIBC.
Note: Data Sets: NCRR Center for Integrative Biomedical Computing (CIBC) data set archive. Download from: http://www.sci.utah.edu/cibc/software.html, 2015.
CIBC.
Note: Cleaver: A MultiMaterial Tetrahedral Meshing Library and Application. Scientific Computing and Imaging Institute (SCI), Download from: http://www.sci.utah.edu/cibc/software.html, 2015.
C.C. Conlin, J.L. Zhang, F. Rousset, C. Vachet, Y. Zhao, K.A. Morton, K. Carlston, G. Gerig, V.S. Lee.
Performance of an Efficient Image-registration Algorithm in Processing MR Renography Data, In J Magnetic Resonance Imaging, July, 2015.
DOI: 10.1002/jmri.25000
PURPOSE:
To evaluate the performance of an edge-based registration technique in correcting for respiratory motion artifacts in magnetic resonance renographic (MRR) data and to examine the efficiency of a semiautomatic software package in processing renographic data from a cohort of clinical patients.
MATERIALS AND METHODS:
The developed software incorporates an image-registration algorithm based on the generalized Hough transform of edge maps. It was used to estimate glomerular filtration rate (GFR), renal plasma flow (RPF), and mean transit time (MTT) from 36 patients who underwent free-breathing MRR at 3T using saturation-recovery turbo-FLASH. The processing time required for each patient was recorded. Renal parameter estimates and model-fitting residues from the software were compared to those from a previously reported technique. Interreader variability in the software was quantified by the standard deviation of parameter estimates among three readers. GFR estimates from our software were also compared to a reference standard from nuclear medicine.
RESULTS:
The time taken to process one patient's data with the software averaged 12 ± 4 minutes. The applied image registration effectively reduced motion artifacts in dynamic images by providing renal tracer-retention curves with significantly smaller fitting residues (P < 0.01) than unregistered data or data registered by the previously reported technique. Interreader variability was less than 10% for all parameters. GFR estimates from the proposed method showed greater concordance with reference values (P < 0.05).
CONCLUSION:
These results suggest that the proposed software can process MRR data efficiently and accurately. Its incorporated registration technique based on the generalized Hough transform effectively reduces respiratory motion artifacts in free-breathing renographic acquisitions. J. Magn. Reson. Imaging 2015.
S. Durrleman, T.P. Fletcher, G. Gerig, M. Niethammer, X. Pennec (Eds.).
Spatio-temporal Image Analysis for Longitudinal and Time-Series Image Data, In Proceedings of the Third International Workshop, STIA 2014, Image Processing, Computer Vision, Pattern Recognition, and Graphics, Vol. 8682, Springer LNCS, 2015.
ISBN: 978-3-319-14905-9
This book constitutes the thoroughly refereed post-conference proceedings of the Third
International Workshop on Spatio-temporal Image Analysis for Longitudinal and Time-
Series Image Data, STIA 2014, held in conjunction with MICCAI 2014 in Boston, MA, USA, in
September 2014.
The 7 papers presented in this volume were carefully reviewed and selected from 15
submissions. They are organized in topical sections named: longitudinal registration and
shape modeling, longitudinal modeling, reconstruction from longitudinal data, and 4D
image processing.
J. Edwards, S. Kumar, V. Pascucci.
Big data from scientific simulations, In Big Data and High Performance Computing, Vol. 26, IOS Press, pp. 32. 2015.
Scientic simulations often generate massive amounts of data used for debugging, restarts, and scientic analysis and discovery. Challenges that practitioners face using these types of big data are unique. Of primary importance is speed of writing data during a simulation, but this need for fast I/O is at odds with other priorities, such as data access time for visualization and analysis, ecient storage, and portability across a variety of supercomputer topologies, congurations, le systems, and storage devices. The computational power of high-performance computing systems continues to increase according to Moore's law, but the same is not true for I/O subsystems, creating a performance gap between computation and I/O. This chapter explores these issues, as well as possible optimization strategies, the use of in situ analytics, and a case study using the PIDX I/O library in a typical simulation.
J. Edwards, E. Daniel, V. Pascucci, C. Bajaj.
Approximating the Generalized Voronoi Diagram of Closely Spaced Objects, In Computer Graphics Forum, Vol. 34, No. 2, Wiley-Blackwell, pp. 299-309. May, 2015.
DOI: 10.1111/cgf.12561
Generalized Voronoi Diagrams (GVDs) have far-reaching applications in robotics, visualization, graphics, and simulation. However, while the ordinary Voronoi Diagram has mature and efficient algorithms for its computation, the GVD is difficult to compute in general, and in fact, has only approximation algorithms for anything but the simplest of datasets. Our work is focused on developing algorithms to compute the GVD efficiently and with bounded error on the most difficult of datasets -- those with objects that are extremely close to each other.
E. Erdil, A.O. Argunsah, T. Tasdizen, D. Unay, M. Cetin.
A joint classification and segmentation approach for dendritic spine segmentation in 2-photon microscopy images, In 2015 IEEE 12th International Symposium on Biomedical Imaging (ISBI), IEEE, April, 2015.
DOI: 10.1109/isbi.2015.7163992
Shape priors have been successfully used in challenging biomedical imaging problems. However when the shape distribution involves multiple shape classes, leading to a multimodal shape density, effective use of shape priors in segmentation becomes more challenging. In such scenarios, knowing the class of the shape can aid the segmentation process, which is of course unknown a priori. In this paper, we propose a joint classification and segmentation approach for dendritic spine segmentation which infers the class of the spine during segmentation and adapts the remaining segmentation process accordingly. We evaluate our proposed approach on 2-photon microscopy images containing dendritic spines and compare its performance quantitatively to an existing approach based on nonparametric shape priors. Both visual and quantitative results demonstrate the effectiveness of our approach in dendritic spine segmentation.
T. Etiene, R.M. Kirby, C. Silva.
An Introduction to Verification of Visualization Techniques, Morgan & Claypool Publishers, 2015.
SCI Institute.
Note: FluoRender: An interactive rendering tool for confocal microscopy data visualization. Scientific Computing and Imaging Institute (SCI) Download from: http://www.fluorender.org, 2015.
Page 40 of 150
