Brain Surface Motion Estimation and Tissue Modeling for Accurate Image Guided Neurosurgery and Surgical Simulation


Table of Contents


Introduction

Image-guided Neurosurgery is a surgical planning technique whereby a volume model of a patient's head is constructed from a set of segmented MRI or CT images and is then registered with the patient's head in the OR using a locating device. Successful guidance depends on the accuracy of the transformation between the graphical and patient spaces, which unfortunately degrades under brain tissue shift and extrusion.


Our Approach to Determining Brain Shift

We propose a method for making the volume account for these factors, by:
Semi-automatic anatomical surface identification. Range-sensing of skin and cortex in the OR. Rigid registration of the skin range image with its MRI/CT homologue, which will serve as a baseline for the next stage.
Non-rigid motion tracking, over time, of the cortical range image. Interpolation of the resulting surface displacement over the whole brain volume, assuming null displacement at the base of the skull, via a realistic finite element model (FEM).

The results shown here were produced by a manual segmentation scheme. Integrating meshing previous to the development of our level-set technique) and rendered with the same colour scheme as the Wayne State head model.

Semi-automatic, Level-Set Based Identification of Anatomical Surfaces in MRI Scan


The Level-Set Based Surface Evolution Model

The skin, cortex, and lesion surface identification technique is an implementation of a recent algorithm, known as a surface evolution model, which imbeds 3D surface segmentation as the zero level-set of a 4D moving front, as illustrated here for the problem of lesser dimensionality, namely identifying 2D contours. This approach possesses certain advantages over physically-based models, in particular its capacity to capture a large variety of topologies and a relative insensitivity to initial conditions.
Illustration of an embedding of 2D contour identification into 3D, a notion which extends to imbedding 3D surfaces into 4D. Skin surface identified from an inward moving front originally initialized as a cube coinciding with the scan perimeter. Cerebral surface identified from an outward moving front originally initialized as a series of small spheres in the white matter.

In general, a surface evolution model is initialized by a user-defined surface completely inside (or outside) the desired anatomical boundary. The front then moves in a strictly outward (inward) manner and features a diffusive term which tends to smooth out the front, a hyperbolic term which pushes the front forward, and two image terms which halt the front at strong gradients. The user-defined surface is processed in the initialization by computing from it a signed distance function, with interior points typically labelled negative, and exterior points positive. Typical results for skin and brain surfaces are shown here. For further information, consult the webpages of either Kaleem Siddiqi, Ron Kimmel, or James Sethian.


Using Fast Marching Level Sets to Compute Distances and Closest Points

The distance function computed from the initial user surface is implemented by a fast marching level-set method which estimates the arrival time of a monotically advancing front. By using a constant unit speed, this arrival time function corresponds to a sub-pixel Euclidean distance map. Next, to reduce the complexity of computing a 4D, rather than 3D, function, we adopt the narrow band approach proposed by
David Adalsteinsson. This technique restricts the computation of the evolution equation to the part of the domain near the zero level-set, or near the evolving 3D surface. This band is also determined by fast marching methods, which amounts to computing a distance function from the evolving 3D surface.

Moreover, the propagative nature of the fast marching algorithm can be exploited to reveal which surface point is closest to a given voxel (i.e.: which surface point possesses the shortest arrival time). This characteristic is useful for the registration later on, as it results in a set of explicit point pairs which lead to a closed-form transformation computation, rather than a lengthy search for optimal transformation parameters which best account for a set of shortest distances. A closest-point map is therefore also computed for each final anatomical surface which we intend to register.


Cortical Surface Displacement Estimation

Acquiring a 3D Cortical Surface

The three-dimensional coordinates of the visible surfaces of the skin and evolving cortex are computed by a commercial range-sensor made by Vitana Corp. (Ottawa, Canada), which uses both laser-based triangulation and defocusing techniques to estimate range. Laser-based range-sensing is the industrial standard for quickly and accurately acquiring dense, quantitative information on 3D shape. The sensor is mounted on a commercial linear positioner, made by Intelligent Actuator. A range image of a human left hemisphere, acquired with this sensor, is featured below.

Laser-based triangulation involves projecting a thin ray of laser light onto an object, capturing the reflected light by a charge-coupled device (CCD), and inferring depth from the pixel position of the reflected point and from the geometric relationship between the position of the laser source and the CCD (as shown in the left image below). Ordinarily, this ray would have to be swept along two axes normal to each other and spanning the anatomical surface of interest. However, our sensor dispenses with one sweep axis by optically spreading the ray into a laser plane. The intersection of this plane with the imaged object instantaneously produces a 2D range-map of a slice of the surface, which appears as a curve of 256 points, or laser profile, on the CCD. A 3D surface is then obtained by linearly sweeping the sensor, ideally in a direction normal to the laser plane, across the surface of interest.
Triangulation geometry. Defocusing geometry.
Range-sensing setup.

In our sensor, defocusing is implemented by allowing the laser light to pass through two holes, at a predetermined distance apart, rather than one. The CCD sees two laser profiles instead of one, and range is determined by measuring the space b (in each pixel column) between the two images of a laser point, as illustrated below, from an inverse relationship between range z and b. The range measurements obtained by triangulation and by defocusing are combined by assigning weights to the two expressions for range. For further detail, consult our SPIE Medical Imaging 1999 paper.

The goal of this procedure is to exploit temporal curvature consistency to produce a displacement map of the cortical surface, as well as the likely area of extrusion (whose curvature is inconsistent), which can then be used to update the equivalent surface on the original graphical model. Furthermore, this surface displacement will be used as a boundary condition for the FEM in order to estimate the volumetric displacement.

Computing a Baseline: Rigid Skin Registration

We estimate cortical surface deformation by first establishing a baseline for this movement, then by registering the initial cortical range image with its homologue identified in the MRI volume, and finally by tracking non-rigid cortical surface motion between time tn and time tn+1. The baseline for cortical movement must be implemented by a rigid registration of the range and MRI-based skin surfaces, which is not a trivial problem because an arbitrarily large transformation must be computed. This problem is generally approached by feature-based techniques in the literature, yet the surfaces involved are potentially featureless, particularly in the back of the head. We are currently working on a calibration procedure, based on the work of Roch Comeau here at the MNI (no website yet), which will allow us to use the patient-MRI space transformation, normally computed for image guidance, to estimate the range-MRI space transformation.

Subsequently, a rigid-body refinement, at t0, computed by an {\it iterative closest point} (ICP) technique (for further detail, consult our forthcoming Medical Image Analysis paper), can bring into precise alignment the range image of the skin and its homolog identified semi-automatically in the MRI scan. Our iterative registration technique is characterized by a distance map being precomputed for the MRI volume from the identified cortical surface, thereby accelerating the computation of distances between closest-point pairs, rather than imposing an expensive search for them. In contrast, closed-form methods for computing transformation parameters all require explicit point-pairs, not distances between them. We make use of the fast marching level-set method to produce a dense map of both closest point labels and the distances to them, allowing us to do away with a search for closest points as well as producing explicit point pairs which can take advantage of closed-form transformation algorithms, thereby making each iteration of our ICP technique essentially instantaneous.


Fast Cortical Surface Tracking: Nonrigid Iterative Closest Point Registration

The non-rigid registration stage uses the same ICP matching technique to produce a smoothed cortical displacement vector function. The expression of the deformation over a 2D domain is a consequence of our explicit displacement information being available only on an open surface, below and outside of which we defer to a realistically valued finite-element model to estimate volumetric movement. Furthermore, in characterizing non-rigid surface movement, we can exploit the regularity of the range domain by using Michael Unser's extremely efficient recursive splines. We adopt a smoothing spline approach, whose first stage is a convolution with a smoothing kernel which produces B-spline coefficients and is followed by a convolution with the indirect transform to yield the smoothed output. The filters can simply be cascaded to implement smoothing in 2D.
Recursive spline block-diagram.

The justification for emphasizing computational efficiency here is two-pronged: clinical acceptability and the inherent temporal underdetermination in estimating non-rigid motion. In order to make the non-rigid motion estimation well-behaved, the smoothing parameter is initially set very high, and is gradually lowered as the process iterates, progressively resolving finer-level motion detail. Just as for the rigid ICP stage, each iteration of the non-rigid surface motion estimation is essentially instantaneous.


Nonrigid Tracking Validation: Elastic Brain Phantom

For the purpose of reproducing non-rigid cortical movement, we have implemented a brain-shaped phantom with elastic material properties. We use a brain-shaped jello mold , into which we pour PolyVinyl Alcohol Cryogel (PVA-C) , as illustrated below. This is a relatively viscous liquid, which upon freezing and thawing, becomes an elastic solid. This phantom rests on a support plate of plexiglass (usable within an MR scanner), and features a moving assembly consisting of a small disk and rod, also of plexiglass, imbedded within the elastic material, as well as some glass beads used to validate the surface displacement/FEM approach to volumetric motion estimation.
PVA-C in liquid state being poured into brain mold. Resulting brain-shaped elastic phantom, after freeze-thaw cycle.

In order to assess non-rigid surface tracking, we turn the set screws under the support plate of our elastic phantom, triggering a deformation of up to 15mm at the top, and image the cortical surface with our range-sensor, while maintaining the support plate fixed with respect to the range-sensor/positioner reference. We then compare the results of a strictly rigid transformation computation with a sequence of rigid and non-rigid ICP registrations, using a rough initial alignment based on the corners of the range domain.
Typical registration results with deformed phantom: MRI level-set surface rigidly registered to the range image of the deformed phantom. MRI level-set surface non-rigidly registered to the range image of the deformed phantom.
Slice through top left image, at y=79.592. Slice through top right image, at y=79.592. Evolution of RMS point-pair distance with ICP iteration.


Tissue Modelling for Volumetric Intra-surgical Motion Estimation and Surgical Simulation

A Generic Head Finite Element Model

The finite element modelling we will adopt for estimating volumetric intrasurgical motion is a static analysis with known boundary displacement on a surface patch of the cortex, and assumed null displacement at the base of the skull. In order to propagate the surface displacement information to the rest of the volume, we will use a FEM whose elements reflect realistic properties of brain tissue (grey and white matter), dura mater, cortico-spinal fluid and bone. Several such models have been published, including those of Albert King et al. of Wayne State University (his student Jesse Ruan's model is shown here) and Mike Miga of Dartmouth College.

Making the Model Patient-Specific

Now in order to make the model adapted to the patient's anatomy, we must construct the elements based on which anatomical surfaces, such as appear on the right, their constituent nodes fall in. This is an instance of the meshing problem. Clearly, we are only interested in tissue types which are identifiable and therefore segmentable in the MRI/CT scan, and especially those which are materially relevant to the simulation.

Transposing the Model to Surgical Simulation

This approach to compensating a graphical model for deformations has another extremely attractive side-benefit: the realistic FEM used to propagate surface information to the rest of the volume can also be adapted to implement a virtual suction aspirator (or a scalpel) as used in the operating room, in the context of surgical simulation. There is research currently being conducted at McGill on haptic feedback and surgical simulation , and a collaboration on patient-specific surgical simulation is foreseen.

Validating the Model

Clinical validation of the cortical surface tracking-FEM approach is planned in the near future, after skin registration and meshing issues are resolved.