Difference between revisions of "ProjectWeek200706:ContrastingTractographyMeasures"

From NAMIC Wiki
Jump to: navigation, search
Line 29: Line 29:
 
<p>3. What statistical methods are most appropriate for quanitfying and testing significance of these assessments? </p>
 
<p>3. What statistical methods are most appropriate for quanitfying and testing significance of these assessments? </p>
  
The answers to these questions will vary depending on the specific algorithm and its application. The group agreed that the best way to proceed was to chose one very specific example that is highly relevant to the NAMIC work to date and focus efforts on that. The methods that arise from this can then be applied to additional areas.
+
The answers to these questions will vary depending on the specific algorithm and its application. The group agreed that the best way to proceed was to chose one very specific example that is highly relevant to the NA-MIC work to date and focus efforts on that. The methods that arise from this can then be applied to additional areas.
  
 
We agreed to begin by studying the results obtained by applying each of the tractography tools to a single dataset and then gathering to present the results to one another and discuss how best to quantify the similarities and differences.  
 
We agreed to begin by studying the results obtained by applying each of the tractography tools to a single dataset and then gathering to present the results to one another and discuss how best to quantify the similarities and differences.  

Revision as of 21:39, 23 June 2007

Home < ProjectWeek200706:ContrastingTractographyMeasures
Arcuate
Cingulum Bundle
Fornix
Internal Capsule
Uncinate Fasciculus


Key Investigators

  • BWH: Marek Kubicki, Doug Markant, Doug Terry, Carl-Fredrik Westin, Katharina Quintus, Jorge Alvarado, Tri Ngo, Sylvain Bouix, Marc Niethammer
  • MGH: Bruce Fischl, Denis Jen
  • Utah: Tom Fletcher, Ross Whitaker, Guido Gerig, Saurav Basu, Davis McKay
  • UNC: Isabelle Corouge, Casey Goodlett, Martin Styner
  • GA Tech: Eric Pichon, John Melonakos, Xavier LaFaucheur, Vandana Mohan, Allen Tannenbaum
  • MIT: Lauren O'Donnell, Polina Golland

Objective

A new initiative has begun in response to a shared vision among Cores 1, 3 and 5 that the field of medical image analysis would be well served by work in the area of validation, calibration and assessment of reliability. Discussions have continued among our participants since then and as a result a plan for the initial work on this front has been articulated.

  • There are many outstanding questions in this domain that we agree are interesting and worth considering such as:

1. What benchmarks should be used to assess performance of a NA-MIC Toolkit algorithm?

2. How can we assess the performance of an algorithm if we have no access to the ground truth of what it is measuring (e.g. the white matter of the brain with tractography)?

3. What statistical methods are most appropriate for quanitfying and testing significance of these assessments?

The answers to these questions will vary depending on the specific algorithm and its application. The group agreed that the best way to proceed was to chose one very specific example that is highly relevant to the NA-MIC work to date and focus efforts on that. The methods that arise from this can then be applied to additional areas.

We agreed to begin by studying the results obtained by applying each of the tractography tools to a single dataset and then gathering to present the results to one another and discuss how best to quantify the similarities and differences.

  • Final goals are:

1. To write up the results for publication with all of us as contributing authors.

2. To make the dataset and our analyzed results available to the scientific community

3. To use our findings to begin to establish benchmarking methods for the NA-MIC toolkit.

Approach

Details:

  • Data sets were provided by Marek Kubicki (put link to descriptor page and download instructions here). N=10 (5 NC, 5 Schiz); each schizophrenic and healthy subjects that are de-identified and not marked as to diagnosis. Each subject will have a 3T high resolution DTI scan, mMRI scan and expert generated Regions of Interest (ROIs) for each subject that are needed for tract definition. Acquisition parameters can be found here: Acquisition Parameters.
  • The tracts to be studied are the cingulum bundle, the uncinate fasiculus, fornix, internal capsule, and the arcuate fasiculus on the left and right sides. To see how the ROI's were defined, click here: ROI Definitions.
  • Each tool developer is responsible for downloading and analyzing the data, optimizing their own algorithm as needed. Keep careful notes on your final processing methods as you will need to teach them to Sonia Pujol who will repeat the analysis independently using all the tools herself for the data to be included in a summary manuscript.
  • Metrics to be collected need to be finalized by this group, but suggestions include measure of FA along the tract, size and/or volume of tract, spatial localization of tract, measure of connectivity. Perhaps also some way to look at group results towards the goal of being able to make statements about differences in health and disease?
  • Participating algorithms (and tester) include:
  1. fiber tracking the UNC way (Guido Gerig or his designee)
  2. Slicer tract tool (Doug Terry/Marek Kubicki)
  3. POI tool (Bruce Fischl/Dennis Jen)
  4. Volumetric connectivity (Ross Whitaker/Tom Fletcher)
  5. Finsler (Allen Tannenbaum/John Melonakos)
  6. Medinrea (PF Filliard)
  7. GTRACT (Vince Magnotta)
  8. Cluster tool (Lauren O'Donnell)
  9. Stochastic Tractography (Tri Ngo/Carl Fredric Westin)
  • We will then hold a 2 day workshop/retreat in Santa Fe, NM. Candidate dates are 2 days during the week of Oct 1-5. That window avoids MICCAI, BIRN, SFN, New Mexico's ballonfest, and the Jewish High Holidays. Please let us know immediately if that week is a NO GO for your group. The workshop/retreat will include a presentation from each group of their results AND a recommendation for how to statstically compare and quantify similarities and differences within and across tools. The outcome of the workshop/retreat will be a set of final agreed upon measurement metrics and method for compare and quantify similarities and differences within and across tools. And the training of Sonia Pujol to run each of the analysis tools.

Progress

Data can be found on BIRN in the following directory:

/home/Projects/NAMIC__0003/Files/PNL/3T_data_sample

with subdirectories:

dwi (raw diffusion data)
dwi-EdCor (raw diffusion data eddy current corrected)
dougt_DTI_ROI (directory of labelmaps)


Additional Information