SCI Publications

2011


G. Gerig, I. Oguz, S. Gouttard, J. Lee, H. An, W. Lin, M. McMurray, K. Grewen, J. Johns, M.A. Styner. “Synergy of image analysis for animal and human neuroimaging supports translational research on drug abuse,” In Frontiers in Child and Neurodevelopmental Psychiatry, Vol. 2, Edited by Linda Mayes, pp. 9 pages. 2011.
ISSN: 1664-0640
DOI: 10.3389/fpsyt.2011.00053



G. Gerig, J.H. Gilmore, W. Lin. “Brain Maturation of Newborns and Infants,” Encyclopedia on Early Childhood Development (online): Brain Development in Children - According to Experts, Montreal, Quebec, Centre of Excellence for Early Childhood Development and Strategic Knowledge Cluster on Early Child Development, pp. 1--6. 2011.

ABSTRACT

Recently, imaging studies of early human development have received more attention, as improved modeling methods might lead to a clearer understanding of the origin, timing, and nature of differences in neurodevelopmental disorders. Non-invasive magnetic resonance imaging (MRI) can provide three-dimensional images of the infant brain in less than 20 minutes, with unprecedented anatomical details and contrast of brain anatomy, cortical and subcortical structures and brain connectivity. Repeating MRI at different stages of development, e.g., in yearly intervals starting after birth, gives scientists the opportunity to study the trajectory of brain growth and compare individual growth trajectories to normative models. These comparisons become highly relevant in personalized medicine, where early diagnosis is a critical juncture for timing and therapy types.



T. Geymayer, A. Lex, M. Streit, D. Schmalstieg. “Visualizing the Effects of Logically Combined Filters,” In Proceedings of the Conference on Information Visualisation (IV '11), IEEE, pp. 47--52. 2011.

ABSTRACT

Filtering data is an essential process in a drill-down analysis of large data sets. Filtering can be necessary for several reasons. The main objective for filters is to uncover the relevant subsets of a dataset. Another, equally relevant goal is to reduce a dataset to dimensions to which either visualization or algorithmic analysis techniques scale. However, with multiple filters applied and possibly even logically combined, it becomes difficult for users to judge the effects of a filter chain. In this paper we present a simple, yet effective way to interactively visualize a sequence of filters and logical combinations of these. Such a visualized filter-pipeline allows analysts to easily judge the effect of every single filter and also their combination on the data set under investigation and therefore, leads to a faster and more efficient workflow.

We also present an implementation of the proposed technique in an information visualization framework for the life sciences. The technique, however, could be employed in many other information visualization contexts as well.



G. Gopalakrishnan, R.M. Kirby, S. Siegel, R. Thakur, W. Gropp, E. Lusk, B.R. de Supinski, M. Schultz, G. Bronevetsky. “Formal Analysis of MPI-Based Parallel Programs: Present and Future,” In Communications of the ACM, pp. (accepted). 2011.



J. Grueninger, H. Hoffman, U. Kloos, J. Krüger. “Multi-Resolution-Display System for Virtual Reality Setups,” In Proceedings of the 14th International Conference on Human-Computer Interaction, HCI International, Lecture Notes in Computer Science, Vol. 6779/2011, pp. 180--189. 2011.
DOI: 10.1007/978-3-642-21716-6_19

ABSTRACT

Most large-area video projection systems offer only limited spacial resolution. Consequently, images of detailed scenery cannot be displayed at full fidelity. A possible but significantly more costly strategy is a tiled projection display. If this solution is not feasible then either aliasing occurs or some anti-aliasing method is used at the cost of reduced scene quality.

In this paper we describe a novel cost effective multi-resolution display system. It allows users to select any part of a stereoscopic projection and view it in significantly higher resolution than possible with the standard projection alone. To achieve this, a pair of video projectors, which can be moved by stepper motors, project a high-resolution inset into a small portion of the low-resolution image. To avoid crosstalk between the low and high resolution projections, a mask is rendered into the low resolution scene to black out the area on the screen that is covered by the inlay.

To demonstrate the effectiveness of our multi-resolution display setup it has been integrated into a number of real life scenarios: a virtual factory, an airplane cabin simulation, and a focus and context volume visualization application (see Figure 1).



Attila Gyulassy, J.A. Levine, V. Pascucci. “Visualization of Discrete Gradient Construction (Multimedia submission),” In Proceedings of the 27th Symposium on Computational Geometry, Paris, France, ACM, pp. 289--290. June, 2011.
DOI: 10.1145/1998196.1998241

ABSTRACT

This video presents a visualization of a recent algorithm to compute discrete gradient fields on regular cell complexes [3]. Discrete gradient fields are used in practical methods that robustly translate smooth Morse theory to combinatorial domains. We describe the stages of the algorithm, highlighting both its simplicity and generality.



L.K. Ha, J. Krüger, J. Comba, S. Joshi, C.T. Silva. “Optimal Multi-Image Processing Streaming Framework on Parallel Heterogeneous Systems,” In Proceedings of Eurographics Symposium on Parallel Graphics and Visualization 2011, Note: Awarded Best Paper!, pp. 1--10. 2011.
DOI: 10.2312/EGPGV/EGPGV11/001-010

ABSTRACT

Atlas construction is an important technique in medical image analysis that plays a central role in understanding the variability of brain anatomy. The construction often requires applying image processing operations to multiple images (often hundreds of volumetric datasets), which is challenging in computational power as well as memory requirements. In this paper we introduce MIP, a Multi-Image Processing streaming framework to harness the processing power of heterogeneous CPU/GPU systems. In MIP we introduce specially designed streaming algorithms and data structures that provides an optimal solution for out-of-core multi-image processing problems both in terms of memory usage and computational efficiency. MIP makes use of the asynchronous execution mechanism supported by parallel heterogeneous systems to efficiently hide the inherent latency of the processing pipeline of out-of-core approaches. Consequently, with computationally intensive problems, the MIP out-of-core solution could achieve the same performance as the in-core solution. We demonstrate the efficiency of the MIP framework on synthetic and real datasets.



L.K. Ha, M.W. Prastawa, G. Gerig, J.H. Gilmore, C.T. Silva. “Efficient Probabilistic and Geometric Anatomical Mapping Using Particle Mesh Approximation on GPUs,” In International Journal of Biomedical Imaging, Special Issue in Parallel Computation in Medical Imaging Applications, Vol. 2011, Note: Article ID 572187, pp. 16 pages. 2011.
DOI: 10.1155/2011/572187



L.K. Ha, J. Krüger, S. Joshi, C.T. Silva. “Multi-scale Unbiased Diffeomorphic Atlas Construction on Multi-GPUs,” Vol. 1, Ch. 10, Morgan Kaufmann, pp. 42. 2011.

ABSTRACT

In this chapter, we present a high performance multi-scale 3D image processing framework to exploit the parallel processing power of multiple graphic processing units (Multi-GPUs) for medical image analysis. We developed GPU algorithms and data structures that can be applied to a wide range of 3D image processing applications and efficiently exploit the computational power and massive bandwidth offered by modern GPUs. Our framework helps scientists solve computationally intensive problems which previously required super computing power. To demonstrate the effectiveness of our framework and to compare to existing techniques, we focus our discussions on atlas construction - the application of understanding the development of the brain and the progression of brain diseases.



X. Hao, R.T. Whitaker, P.T. Fletcher. “Adaptive Riemannian Metrics for Improved Geodesic Tracking of White Matter,” In Information Processing in Medical Imaging (IPMI), Lecture Notes in Computer Science (LNCS), Vol. 6801/2011, pp. 13--24. 2011.
DOI: 10.1007/978-3-642-22092-0_2



D.E. Hart, M. Berzins, C.E. Goodyer, P.K. Jimack. “Using Adjoint Error Estimation Techniques for Elastohydrodynamic Lubrication Line Contact Problems,” In International Journal for Numerical Methods in Fluids, Vol. 67, Note: Published online 29 October, pp. 1559--1570. 2011.



H.C. Hazlett, M. Poe, G. Gerig, M. Styner, C. Chappell, R.G. Smith, C. Vachet, J. Piven. “Early Brain Overgrowth in Autism Associated with an Increase in Cortical Surface Area Before Age 2,” In Arch of Gen Psych, Vol. 68, No. 5, pp. 467--476. 2011.
DOI: 10.1001/archgenpsychiatry.2011.39



C.R. Henak, B.J. Ellis, M.D. Harris, A.E. Anderson, C.L. Peters, J.A. Weiss. “Role of the acetabular labrum in load support across the hip joint,” In Journal of Biomechanics, Vol. 44, No. 12, pp. 2201-2206. 2011.

ABSTRACT

The relatively high incidence of labral tears among patients presenting with hip pain suggests that the acetabular labrum is often subjected to injurious loading in vivo. However, it is unclear whether the labrum participates in load transfer across the joint during activities of daily living. This study examined the role of the acetabular labrum in load transfer for hips with normal acetabular geometry and acetabular dysplasia using subject-specific finite element analysis. Models were generated from volumetric CT data and analyzed with and without the labrum during activities of daily living. The labrum in the dysplastic model supported 4–11\% of the total load transferred across the joint, while the labrum in the normal model supported only 1–2\% of the total load. Despite the increased load transferred to the acetabular cartilage in simulations without the labrum, there were minimal differences in cartilage contact stresses. This was because the load supported by the cartilage correlated with the cartilage contact area. A higher percentage of load was transferred to the labrum in the dysplastic model because the femoral head achieved equilibrium near the lateral edge of the acetabulum. The results of this study suggest that the labrum plays a larger role in load transfer and joint stability in hips with acetabular dysplasia than in hips with normal acetabular geometry.



L. Hogrebe, A. Paiva, E. Jurrus, C. Christensen, M. Bridge, J.R. Korenberg, T. Tasdizen. “Trace Driven Registration of Neuron Confocal Microscopy Stacks,” In IEEE International Symposium on Biomedical Imaging (ISBI), pp. 1345--1348. 2011.
DOI: 10.1109/ISBI.2011.5872649



A. Irimia, M.C. Chambers, J.R. Alger, M. Filippou, M.W. Prastawa, Bo Wang, D. Hovda, G. Gerig, A.W. Toga, R. Kikinis, P.M. Vespa, J.D. Van Horn. “Comparison of acute and chronic traumatic brain injury using semi-automatic multimodal segmentation of MR volumes,” In Journal of Neurotrauma, Vol. 28, No. 11, pp. 2287--2306. November, 2011.
DOI: 10.1089/neu.2011.1920
PubMed ID: 21787171

ABSTRACT

Although neuroimaging is essential for prompt and proper management of traumatic brain injury (TBI), there is a regrettable and acute lack of robust methods for the visualization and assessment of TBI pathophysiology, especially for of the purpose of improving clinical outcome metrics. Until now, the application of automatic segmentation algorithms to TBI in a clinical setting has remained an elusive goal because existing methods have, for the most part, been insufficiently robust to faithfully capture TBI-related changes in brain anatomy. This article introduces and illustrates the combined use of multimodal TBI segmentation and time point comparison using 3D Slicer, a widely-used software environment whose TBI data processing solutions are openly available. For three representative TBI cases, semi-automatic tissue classification and 3D model generation are performed to perform intra-patient time point comparison of TBI using multimodal volumetrics and clinical atrophy measures. Identification and quantitative assessment of extra- and intra-cortical bleeding, lesions, edema, and diffuse axonal injury are demonstrated. The proposed tools allow cross-correlation of multimodal metrics from structural imaging (e.g., structural volume, atrophy measurements) with clinical outcome variables and other potential factors predictive of recovery. In addition, the workflows described are suitable for TBI clinical practice and patient monitoring, particularly for assessing damage extent and for the measurement of neuroanatomical change over time. With knowledge of general location, extent, and degree of change, such metrics can be associated with clinical measures and subsequently used to suggest viable treatment options.

Keywords: namic



B.M. Isaacson, J.G. Stinstra, R.D. Bloebaum, COL P.F. Pasquina, R.S. MacLeod. “Establishing Multiscale Models for Simulating Whole Limb Estimates of Electric Fields for Osseointegrated Implants,” In IEEE Transactions on Biomedical Engineering, Vol. 58, No. 10, pp. 2991--2994. 2011.
DOI: 10.1109/TBME.2011.2160722
PubMed ID: 21712151
PubMed Central ID: PMC3179554

ABSTRACT

Although the survival rates of warfighters in recent conflicts are among the highest in military history, those who have sustained proximal limb amputations may present additional rehabilitation challenges. In some of these cases, traditional prosthetic limbs may not provide adequate function for service members returning to an active lifestyle. Osseointegration has emerged as an acknowledged treatment for those with limited residual limb length and those with skin issues associated with a socket together. Using this technology, direct skeletal attachment occurs between a transcutaneous osseointegrated implant (TOI) and the host bone, thereby eliminating the need for a socket. While reports from the first 100 patients with a TOI have been promising, some rehabilitation regimens require 12-18 months of restricted weight bearing to prevent overloading at the bone-implant interface. Electrically induced osseointegration has been proposed as an option for expediting periprosthetic fixation and preliminary studies have demonstrated the feasibility of adapting the TOI into a functional cathode. To assure safe and effective electric fields that are conducive for osseoinduction and osseointegration, we have developed multiscale modeling approaches to simulate the expected electric metrics at the bone--implant interface. We have used computed tomography scans and volume segmentation tools to create anatomically accurate models that clearly distinguish tissue parameters and serve as the basis for finite element analysis. This translational computational biological process has supported biomedical electrode design, implant placement, and experiments to date have demonstrated the clinical feasibility of electrically induced osseointegration.



S.A. Isaacson, R.M. Kirby. “Numerical Solution of Linear Volterra Integral Equations of the Second Kind with Sharp Gradients,” In Journal of Computational and Applied Mathematics, Vol. 235, No. 14, pp. 4283--4301. 2011.

ABSTRACT

Collocation methods are a well-developed approach for the numerical solution of smooth and weakly singular Volterra integral equations. In this paper, we extend these methods through the use of partitioned quadrature based on the qualocation framework, to allow the efficient numerical solution of linear, scalar Volterra integral equations of the second kind with smooth kernels containing sharp gradients. In this case, the standard collocation methods may lose computational efficiency despite the smoothness of the kernel. We illustrate how the qualocation framework can allow one to focus computational effort where necessary through improved quadrature approximations, while keeping the solution approximation fixed. The computational performance improvement introduced by our new method is examined through several test examples. The final example we consider is the original problem that motivated this work: the problem of calculating the probability density associated with a continuous-time random walk in three dimensions that may be killed at a fixed lattice site. To demonstrate how separating the solution approximation from quadrature approximation may improve computational performance, we also compare our new method to several existing Gregory, Sinc, and global spectral methods, where quadrature approximation and solution approximation are coupled.



T. Ize, C.D. Hansen. “RTSAH Traversal Order for Occlusion Rays,” In Computer Graphics Forum, Vol. 30, No. 2, Wiley-Blackwell, pp. 297--305. April, 2011.
DOI: 10.1111/j.1467-8659.2011.01861.x

ABSTRACT

We accelerate the finding of occluders in tree based acceleration structures, such as a packetized BVH and a single ray kd-tree, by deriving the ray termination surface area heuristic (RTSAH) cost model for traversing an occlusion ray through a tree and then using the RTSAH to determine which child node a ray should traverse first instead of the traditional choice of traversing the near node before the far node. We further extend RTSAH to handle materials that attenuate light instead of fully occluding it, so that we can avoid superfluous intersections with partially transparent objects. For scenes with high occlusion, we substantially lower the number of traversal steps and intersection tests and achieve up to 2x speedups.



T. Ize, C. Brownlee, C.D. Hansen. “Real-Time Ray Tracer for Visualizing Massive Models on a Cluster,” In Proceedings of the 2011 Eurographics Symposium on Parallel Graphics and Visualization, pp. 61--69. 2011.

ABSTRACT

We present a state of the art read-only distributed shared memory (DSM) ray tracer capable of fully utilizing modern cluster hardware to render massive out-of-core polygonal models at real-time frame rates. Achieving this required adapting a state of the art packetized BVH acceleration structure for use with DSM and modifying the mesh and BVH data layouts to minimize communication costs. Furthermore, several design decisions and optimizations were made to take advantage of InfiniBand interconnects and multi-core machines.



S. Jadhav, H. Bhatia, P.-T. Bremer, J.A. Levine, L.G. Nonato, V. Pascucci. “Consistent Approximation of Local Flow Behavior for 2D Vector Fields,” In Mathematics and Visualization, Springer, pp. 141--159. Nov, 2011.
DOI: 10.1007/978-3-642-23175-9 10

ABSTRACT

Typically, vector fields are stored as a set of sample vectors at discrete locations. Vector values at unsampled points are defined by interpolating some subset of the known sample values. In this work, we consider two-dimensional domains represented as triangular meshes with samples at all vertices, and vector values on the interior of each triangle are computed by piecewise linear interpolation.

Many of the commonly used techniques for studying properties of the vector field require integration techniques that are prone to inconsistent results. Analysis based on such inconsistent results may lead to incorrect conclusions about the data. For example, vector field visualization techniques integrate the paths of massless particles (streamlines) in the flow or advect a texture using line integral convolution (LIC). Techniques like computation of the topological skeleton of a vector field, require integrating separatrices, which are streamlines that asymptotically bound regions where the flow behaves differently. Since these integrations may lead to compound numerical errors, the computed streamlines may intersect, violating some of their fundamental properties such as being pairwise disjoint. Detecting these computational artifacts to allow further analysis to proceed normally remains a significant challenge.