<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://www.na-mic.org/w/index.php?action=history&amp;feed=atom&amp;title=FBIRN%3ANeuroInfStatsTuesAM2006</id>
	<title>FBIRN:NeuroInfStatsTuesAM2006 - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://www.na-mic.org/w/index.php?action=history&amp;feed=atom&amp;title=FBIRN%3ANeuroInfStatsTuesAM2006"/>
	<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=FBIRN:NeuroInfStatsTuesAM2006&amp;action=history"/>
	<updated>2026-04-17T12:12:05Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.33.0</generator>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=FBIRN:NeuroInfStatsTuesAM2006&amp;diff=3626&amp;oldid=prev</id>
		<title>Andy: Update from Wiki</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=FBIRN:NeuroInfStatsTuesAM2006&amp;diff=3626&amp;oldid=prev"/>
		<updated>2006-12-18T13:26:28Z</updated>

		<summary type="html">&lt;p&gt;Update from Wiki&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;= Data Mining =&lt;br /&gt;
&lt;br /&gt;
* History of Data Mining&lt;br /&gt;
** 1960's Tukey&lt;br /&gt;
** Predictive Modeling (CART)&lt;br /&gt;
** Machine Learning (CS)&lt;br /&gt;
** Relational Databases&lt;br /&gt;
** Web&lt;br /&gt;
** Hardware&lt;br /&gt;
* Convergence of above technologies has lead to data mining&lt;br /&gt;
** There is a lot of hype regarding data mining&lt;br /&gt;
* FBIRN has a small N in the data mining sense&lt;br /&gt;
** Clustering, look for subgroups in the schizophrenia patients&lt;br /&gt;
** fBIRN data has a rich structure&lt;br /&gt;
** Focused on data analysis right now, do algorithm comparisons&lt;br /&gt;
** Make sure that the raw spatial data, timecourse data is available&lt;br /&gt;
* Standardized data for data analyses&lt;br /&gt;
** MRI Machine --&amp;gt; &amp;quot;raw data&amp;quot; --&amp;gt; intermediate data --&amp;gt; Beta-maps, etc&lt;br /&gt;
*** inter-subject issues in preprocessing&lt;br /&gt;
* data analysis recommendations&lt;br /&gt;
** give someting standard and consistent across sites&lt;br /&gt;
** should data vetting be performed, runs or subjects tossed based on results of first level analysis&lt;br /&gt;
** time series of variance, auto-correlation&lt;br /&gt;
** having standard datasets available so other disciplines (e.g., computer scientists, statisticians) can look at the data&lt;br /&gt;
** proposal, FIPS 1.0 produces standard results (beta-maps) on phaseII, then spend 6 months improving the pipeline&lt;br /&gt;
** make data available after each major preprocessing step&lt;br /&gt;
*** after pre-whitening/smoothing&lt;br /&gt;
*** after motion correction (would need to modify FIPS/FEAT)&lt;br /&gt;
*** after slice timing correction (would need to modify FIPS/FEAT)&lt;br /&gt;
*** currently FIPS final products create the output of FEAT&lt;br /&gt;
** phaseII&lt;br /&gt;
*** uncompressed raw data is just over 1Gb/subject visit&lt;br /&gt;
*** intermediate time series 20Gb/subject visit&lt;br /&gt;
*** may be problematic to store all of the intermediate steps, due to disk space and downloading times&lt;br /&gt;
* what can we do for every subject?&lt;br /&gt;
* what can we do on a subset of subjects?&lt;br /&gt;
* FIAC&lt;br /&gt;
** standard data products&lt;br /&gt;
** 1st level&lt;br /&gt;
*** raw&lt;br /&gt;
*** motion corrected&lt;br /&gt;
**** motion parameters&lt;br /&gt;
*** slice-time corrected&lt;br /&gt;
*** meta-data, details of the paradigm&lt;br /&gt;
**** best practices, what we think the design matrix should be&lt;br /&gt;
**** what happened at what time, how long the blocks were, etc.&lt;br /&gt;
** 2nd level&lt;br /&gt;
*** constrast copes in standard space&lt;br /&gt;
*** varcopes in standard space&lt;br /&gt;
*** meta-data,&lt;br /&gt;
**** what effect&lt;br /&gt;
**** degrees of freedom&lt;br /&gt;
**** smoothness&lt;br /&gt;
**** T image&lt;br /&gt;
**** threshold&lt;br /&gt;
* Standard Data Products Plan&lt;br /&gt;
** Establish a working group&lt;br /&gt;
** Establish the requirements of what is released besides raw data and when&lt;br /&gt;
** QA reports on raw and on intermediate data products&lt;br /&gt;
* Data QA&lt;br /&gt;
** Variance of image over time series&lt;br /&gt;
** Variacne oc (Xt-Xt+1) over time series&lt;br /&gt;
*** scale median to 100&lt;br /&gt;
** Good for residuals&lt;br /&gt;
*** (Outlier count per image / Expected outlier count ) * 100 for each image yields a time series&lt;br /&gt;
*** Normality test at each voxel, look at how many p values smaller than 0.05&lt;br /&gt;
**** (Number of significant voxels / Expected significant voxels) * 100&lt;/div&gt;</summary>
		<author><name>Andy</name></author>
		
	</entry>
</feed>