Algorithm:GATech:DWMRI Musings

From NAMIC Wiki
Revision as of 21:49, 25 July 2007 by Melonakos (talk | contribs)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search
Home < Algorithm:GATech:DWMRI Musings

DW-MRI Musings

Back to Georgia Tech DW-MRI Geodesic Active Contours.

This page contains a hodge podge of thoughts on Diffusion-Weighted MRI (DW-MRI) processing. From a bird's eye view, the most general question for a DW-MRI data processor is: What clinically relevant information can be leveraged from DW-MRI data and used for scientific and clinical studies? Furthermore, what algorithms, programming constructs, clinical guidance, and training are needed to make this happen? The following are thoughts that may be considered when attempting to answer these questions, in no particular order:

DW-MRI vs DT-MRI In answering the questions posed above, it is common for people to assume from the onset that we are using terminology related to tensors (i.e. DT-MRI, DTI). The raw data coming out of the scanner is referred to as DW-MRI or DWI. The difference between the "T" and the "W" is that the "T" has undergone a preprocessing step which computes tensors at each voxel which are a function of the original "W" data. Hence, the data resulting from this preprocessing step is referred to at DT-MRI. There are advantages and disadvantages to using this preprocessing tensor construction step. The point here is that when we think about trying to find ways to go from the scanner data to clinical answers, we are talking about going from DW-MRI data to clinical answers. In other words, it is okay to think outside the box and to consider ways in which ellipsoidal constraints arising from the tensor model may be meaningfully relaxed.

Validation Soap Box When we consider assessing the utility of a particular tool in serving the needs of our clinical customers, we should stop and consider the factors upon which our assessment is based. This is particular important when attempting to validate results in an environment absent of ground truth. In the absence of ground truth, tools are judged according to: visual appeal, reproducibility, user-friendliness, and (if you're shooting for the stars and want something for which there is ground truth) accuracy in diagnosis.

Broadly speaking, each tool consists of four components which contribute toward the utility of the tool: the algorithm, the engineering, the clinical knowledge, and the training.