<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://www.na-mic.org/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Aris</id>
	<title>NAMIC Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://www.na-mic.org/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Aris"/>
	<link rel="alternate" type="text/html" href="https://www.na-mic.org/wiki/Special:Contributions/Aris"/>
	<updated>2026-04-03T22:31:56Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.33.0</generator>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:Meeting_Minutes_20090320&amp;diff=35770</id>
		<title>SDIWG:Meeting Minutes 20090320</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:Meeting_Minutes_20090320&amp;diff=35770"/>
		<updated>2009-03-23T04:30:04Z</updated>

		<summary type="html">&lt;p&gt;Aris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==NCBC Joint Working Group Meeting==&lt;br /&gt;
&lt;br /&gt;
[[SDIWG:Software_and_Data_Integration_Working_Group|Top page of SDIWG web site]]&lt;br /&gt;
&lt;br /&gt;
===Notes===&lt;br /&gt;
''Friday March 20, 2009: 2:30-3:30 PM EST''&amp;lt;br&amp;gt;&lt;br /&gt;
''Please contact Peter Lyster for information [mailto:lysterp@mail.nih.gov Peter Lyster]''&amp;lt;br&amp;gt;&lt;br /&gt;
''[[SDIWG:Software_Engineering_Body_of_Knowledge_Across_NCBC_Biocomputing_Centers | SDIWG Meetings Page]]''&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;b&amp;gt;Save the date for the next SDWIG tcon/Connect, three months from now: Friday June 20, 2009&amp;lt;/b&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Tcon Agenda===&lt;br /&gt;
&lt;br /&gt;
*Roll Call/Note-taker (Aris Floratos) (5 min)&lt;br /&gt;
*Minutes from most recent SDIWG tcon [[SDIWG:Meeting_Minutes_20080919|Sepember 19, 2008]] (5 min)&lt;br /&gt;
*Status check on '''three working groups''': &lt;br /&gt;
**[[SDIWG:_NCBC_Resource_Yellow_Pages_and_Software_Ontologies|Biositemaps Resourceome]]&lt;br /&gt;
***State of [http://www.biositemap.org Biositemaps].  The underlying technology for Biositemaps is the [http://bioportal.bioontology.org/visualize/39521 Biomedical Resource Ontology (BRO) V2.6] and the [http://biositemaps.bioontology.org/editor/ Information Model (IM) which is encapsulated in the Biositemaps Editor. Please try it out.]&lt;br /&gt;
*** The [[Tiger_team_biositemaps | &amp;lt;b&amp;gt; Tiger Team&amp;lt;/b&amp;gt;]] is working on this phase of biositemaps.&lt;br /&gt;
*** If you want to view the working group thread, suggest you start at the [[SDIWG:Meeting_Minutes_20080716 | Four basic activities, Authoring/Consuming of Biositemaps/BRO]]&lt;br /&gt;
*** [http://www.na-mic.org/Wiki/index.php/SDIWG:Action_Items_20090212  Current list of Action Items and Feature Requests for Tiger Team]&lt;br /&gt;
*** [[Plans for adoption within and beyond NCBCs | Plans for adoption within and beyond NCBCs]]&lt;br /&gt;
***[[Media:Biositemaps_white_paper_v4.0.doc | Biositemaps white paper v4.0]] (Needs Updating: Peter, Ivo, Daniel).&lt;br /&gt;
***[[Cheat Sheet for NCBC Leads to adopt Biositemaps | Cheat Sheet for NCBC Leads to adopt Biositemaps]]&lt;br /&gt;
** [[SDIWG:_NCBC_Scientific_Ontologies|Scientific Ontologies]]&lt;br /&gt;
** [http://www.ncbcs.org/dbp_interactions.html NCBC Driving Biological Projects Interactions and Impact]&lt;br /&gt;
**NCBC Dissemination and shared calendar: Dissemination page is at  http://wiki.simtk.org/ncbcdisseminate/. The prototype calendar is at http://www.ncbcs.org/cal.html with RSS feeds available for each center at http://www.ncbcs.org/summary-cal.html.  Please check it out and provide feedback. (Joy)&lt;br /&gt;
**Meeting on Shared Names URI April 29-30 Boston (Mark)&lt;br /&gt;
**Brief update on NCBC supplements: Biositemaps (Ivo, Beth, Mark); Image data (Will); …&lt;br /&gt;
**Simbios Glossy on working groups (Joy, Brian)&lt;br /&gt;
* Questions/Information Sharing&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Minutes===&lt;br /&gt;
* '''Note taker''': [Suggested order of note takers for future meetings: '''Aris Floratos (this one)'''; Schroeder, Joy Ku; Rubin; Jags, Dinov, Murphy]&lt;br /&gt;
* Review of [[SDIWG:Meeting_Minutes_20080919|September 19, 2008 Minutes]] &lt;br /&gt;
* Attendees: &lt;br /&gt;
&lt;br /&gt;
'''''Karin Remington - Update on Common Fund'''''&lt;br /&gt;
---------------------------------------&lt;br /&gt;
* Majority of common fund reserved for new submissions. Administrative/competitive revision supplements will be handled on an ad-hoc basis, a smaller portion of the common fund will go there. This in addition to the stimulus $$$ to be handed out by individual ICs. &lt;br /&gt;
* In response to a question by Andrea Califano if one could submit a supplement for 50% of a Center's budget, Karin said that she will inquire about specifics but that 50% seemed rather on the high side of what would/could be supported. The point was also made that there is a single pool of money for both supplement types (admin/competitive) and that competitive supplements are appropriate only in case of expanding work scope. &lt;br /&gt;
*In terms of timelines, there is a good chance that there will be additional calls or supplements, in addition to the one in the recent RFA. Centers that wish to pursue supplements should discuss their proposals with their program officer, who needs to sign off to supplement requests. In the context of such conversations it is appropriate to inquire if the program officer's IC would be willing to pick up the supplemet funding (instead of having the $$$ come from the Common Fund).&lt;br /&gt;
* Regarding the NCBC re-competition, the application submission deadline will probably be sometime in September-October.&lt;br /&gt;
&lt;br /&gt;
'''''Andrea Califano - Update on DBP WG'''''&lt;br /&gt;
----------------------------------&lt;br /&gt;
* Update of WG calls. Agreed to create an interactive DBP-ome. Provide info on projects, tools, publications, interactions, etc. Points of contact have been defined at the various Centers to collect data and develop sw. A web site (similar to that developed by the yellow-pages WG) will be created soon.&lt;br /&gt;
* Another initiative agreed upon is to have a web site to publish movies of seminars that take place in the various institutes. Lyster/Athey commented that the site could also be used to post other related talks, e.g., those from Valerie Florance's seminars series.&lt;br /&gt;
* (Cavacoli). He sent a message earlier today about the info needed for the DBP-ome. People are responding rapidly. He expects him and Zhong Li to start working on the collected data. Also, mentioned that NCIBI has a software for sharing seminar movies, we can maybe use that. &lt;br /&gt;
&lt;br /&gt;
'''''Beth Kirschner - Biosite maps'''''&lt;br /&gt;
-------------------&lt;br /&gt;
* The biosite map site is complete. All needed search/browsing functionality is complete. Next step is populating with content about various Center tools.&lt;br /&gt;
* (Athey): can this be used to catalogue the software tools in the 38 CTSAs (Besich and him are spearheading)? Beth said yes, and that they are working on this with the CTSA program.&lt;br /&gt;
* (Lyster): a number of other NIH programs are interested in adopting the biosite maps. He said he will be contacting Center contacts to initiate the process of adding content.&lt;br /&gt;
&lt;br /&gt;
'''''Scientific Ontologies'''''&lt;br /&gt;
---------------------&lt;br /&gt;
* (Athey): The bioontologies draft has been sent to Mark Musen who plans to review it shortly.&lt;br /&gt;
&lt;br /&gt;
'''''Joy Ku - Dissemination'''''&lt;br /&gt;
----------------------&lt;br /&gt;
* Talked about the calendar on the ncbcs.org site, listing the various Center events. Described the Web GUI and how it can be used to acquire information about the various events. The idea is that each Center will appoint a contact person that will be responsible for updating their portion of the calendar.&lt;br /&gt;
&lt;br /&gt;
'''''Misc'''''&lt;br /&gt;
----&lt;br /&gt;
* (Musen): talked about an effort spearheaded in coordination with W3 for defining standard terms that can be used to annotate medical info on the web. One day workshop on April 29 (pior to the BioIT meeting) to look at this issue.&lt;br /&gt;
&lt;br /&gt;
* (Lyster): also mentioned a couple of other meetings related to bioontologies. There was a discussion if the ncbcs.org calendar could use a &amp;quot;general purpose&amp;quot; category to capture such interesting but non-Center specific events.&lt;br /&gt;
&lt;br /&gt;
* (Ku): there is a Symbios page that lists notes from a recent meeting with NCBO, where they discussed efforts for coordinating dissemination efforrts. There was general agreement for an effort similar to the DBP group, where Centers work together to promote the coordinated dissemination of the NCBC products.&lt;br /&gt;
&lt;br /&gt;
===Action Items===&lt;br /&gt;
* TBD&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10631</id>
		<title>SDIWG:Software Engineering At MAGNet</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10631"/>
		<updated>2007-05-19T22:38:07Z</updated>

		<summary type="html">&lt;p&gt;Aris: /* Contact */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__TOC__&lt;br /&gt;
==Process==&lt;br /&gt;
The software development process in MAGNet typically takes place in 2 phases:&lt;br /&gt;
* Prototype tools are developed and field-tested at various investigator labs in the context of Core 1, 2 and 3 activities. These tools are usually command line versions, coded in an assortment of languages (C, C++, Perl, etc) and/or computational environments (MATLAB, R). In the majority of cases, analysis results are represented as text files that are either inspected with text editors or are transformed (through custom programs) to properly formatted inputs for downstream analysis/visualization tools.&lt;br /&gt;
* When prototypes have reached a satisfactory level of maturity the software engineering group takes over the task of integrating them into the Center's bioinformatics platform, geWorkbench, http://www.geworkbench.org. Operationally, each integration activity is carried out following a number of steps deriving from a standard software life cycle process:&lt;br /&gt;
:# '''Collection of business requirements''': an initial meeting is arranged with the investigator whose prototype is slated for integration. During the meeting high level requirements are collected, outlining the intended functionality of the integrated component. Such a meeting usually involves (i) a demo of geWorkbench, (ii) an overview presentation outlining the science behind the investigator's prototype, (iii) a demo of the prototype, and (iv) a brainstorming session aimed at identifying the most meaningful manner to integrate the prototype into geWorkbench so that the prototype leverages as much as possible of the functionality already present in geWorkbench.&lt;br /&gt;
:# '''Development of Use case document''':  based on the information collected at the investigator meeting a detailed Use Case requirements document is developed which describes precisely what the user interface will look like and how users will interact with it. The Use Case document is subject to review and approval by the investigator. An example of such a document can be found [[Media:CNKB_UC.pdf | here]].&lt;br /&gt;
:# '''Development of user interface/computational service''': based on the requirements laid out in the use case document, the development team (i) develops the user interface for the integrated version of the investigator tool, and (ii) ports any associated compute server to the grid-enabled server framework used by geWorkbench (which is based on the caGrid software, https://cabig.nci.nih.gov/workspaces/Architecture/caGrid, a grid middleware layer developed in the context of the caBIG initiative). Unit tests are an integral part of the development process and are run automatically on a nightly basis.&lt;br /&gt;
:# '''System Testing''': After the prototype in integrated, an integration and system testing cycle is performed. Our group uses a formal testing process based on detailed test scripts whose execution status is tracked in a custom database we have developed for this purpose. In this database, failed scripts are linked to their associated defect reports within our defect management system, http://wiki.c2b2.columbia.edu/mantis/.&lt;br /&gt;
:# '''Documentation''': The final step in the development process is the generation of detailed end-user documentation, in the form of on-line help, tutorials, end-user guides and training slides. Links to these materials are available at: http://wiki.c2b2.columbia.edu/workbench/index.php/Project_Documentation. &lt;br /&gt;
&lt;br /&gt;
geWorkbench is an open-source platform and we encourage and welcome code contributions by the community. To facilitate such contributions we have made the entire code base freely available to everyone through our community development project site, http://gforge.nci.nih.gov/projects/geworkbench/, registered with the NCI GForge project. This site enables us to leverage the infrastructure support that GForge offers to collaborative development projects, including access to a CVS server, streamlined setup of mailing lists and user-forums, posting software releases, etc.&lt;br /&gt;
&lt;br /&gt;
==License==&lt;br /&gt;
At present, geWorkbench in made available under the licensing terms stated at http://wiki.c2b2.columbia.edu/workbench/index.php/GeWorkbench_License. We are currently in the process of modularizing our licensing , in order to accommodate the needs of different labs/tools.&lt;br /&gt;
&lt;br /&gt;
==Contact==&lt;br /&gt;
The software engineering group at MAGNet is managed by Aris Floratos (&amp;amp;#102;&amp;amp;#108;&amp;amp;#111;&amp;amp;#114;&amp;amp;#97;&amp;amp;#116;&amp;amp;#111;&amp;amp;#115;&amp;amp;#64;&amp;amp;#99;&amp;amp;#50;&amp;amp;#98;&amp;amp;#50;&amp;amp;#46;&amp;amp;#99;&amp;amp;#111;&amp;amp;#108;&amp;amp;#117;&amp;amp;#109;&amp;amp;#98;&amp;amp;#105;&amp;amp;#97;&amp;amp;#46;&amp;amp;#101;&amp;amp;#100;&amp;amp;#117;).&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10630</id>
		<title>SDIWG:Software Engineering At MAGNet</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10630"/>
		<updated>2007-05-19T22:35:51Z</updated>

		<summary type="html">&lt;p&gt;Aris: /* Contact */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__TOC__&lt;br /&gt;
==Process==&lt;br /&gt;
The software development process in MAGNet typically takes place in 2 phases:&lt;br /&gt;
* Prototype tools are developed and field-tested at various investigator labs in the context of Core 1, 2 and 3 activities. These tools are usually command line versions, coded in an assortment of languages (C, C++, Perl, etc) and/or computational environments (MATLAB, R). In the majority of cases, analysis results are represented as text files that are either inspected with text editors or are transformed (through custom programs) to properly formatted inputs for downstream analysis/visualization tools.&lt;br /&gt;
* When prototypes have reached a satisfactory level of maturity the software engineering group takes over the task of integrating them into the Center's bioinformatics platform, geWorkbench, http://www.geworkbench.org. Operationally, each integration activity is carried out following a number of steps deriving from a standard software life cycle process:&lt;br /&gt;
:# '''Collection of business requirements''': an initial meeting is arranged with the investigator whose prototype is slated for integration. During the meeting high level requirements are collected, outlining the intended functionality of the integrated component. Such a meeting usually involves (i) a demo of geWorkbench, (ii) an overview presentation outlining the science behind the investigator's prototype, (iii) a demo of the prototype, and (iv) a brainstorming session aimed at identifying the most meaningful manner to integrate the prototype into geWorkbench so that the prototype leverages as much as possible of the functionality already present in geWorkbench.&lt;br /&gt;
:# '''Development of Use case document''':  based on the information collected at the investigator meeting a detailed Use Case requirements document is developed which describes precisely what the user interface will look like and how users will interact with it. The Use Case document is subject to review and approval by the investigator. An example of such a document can be found [[Media:CNKB_UC.pdf | here]].&lt;br /&gt;
:# '''Development of user interface/computational service''': based on the requirements laid out in the use case document, the development team (i) develops the user interface for the integrated version of the investigator tool, and (ii) ports any associated compute server to the grid-enabled server framework used by geWorkbench (which is based on the caGrid software, https://cabig.nci.nih.gov/workspaces/Architecture/caGrid, a grid middleware layer developed in the context of the caBIG initiative). Unit tests are an integral part of the development process and are run automatically on a nightly basis.&lt;br /&gt;
:# '''System Testing''': After the prototype in integrated, an integration and system testing cycle is performed. Our group uses a formal testing process based on detailed test scripts whose execution status is tracked in a custom database we have developed for this purpose. In this database, failed scripts are linked to their associated defect reports within our defect management system, http://wiki.c2b2.columbia.edu/mantis/.&lt;br /&gt;
:# '''Documentation''': The final step in the development process is the generation of detailed end-user documentation, in the form of on-line help, tutorials, end-user guides and training slides. Links to these materials are available at: http://wiki.c2b2.columbia.edu/workbench/index.php/Project_Documentation. &lt;br /&gt;
&lt;br /&gt;
geWorkbench is an open-source platform and we encourage and welcome code contributions by the community. To facilitate such contributions we have made the entire code base freely available to everyone through our community development project site, http://gforge.nci.nih.gov/projects/geworkbench/, registered with the NCI GForge project. This site enables us to leverage the infrastructure support that GForge offers to collaborative development projects, including access to a CVS server, streamlined setup of mailing lists and user-forums, posting software releases, etc.&lt;br /&gt;
&lt;br /&gt;
==License==&lt;br /&gt;
At present, geWorkbench in made available under the licensing terms stated at http://wiki.c2b2.columbia.edu/workbench/index.php/GeWorkbench_License. We are currently in the process of modularizing our licensing , in order to accommodate the needs of different labs/tools.&lt;br /&gt;
&lt;br /&gt;
==Contact==&lt;br /&gt;
The software engineering group at MAGNet is managed by Aris Floratos (floratos at c2b2 dot columbia dot edu).&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10629</id>
		<title>SDIWG:Software Engineering At MAGNet</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10629"/>
		<updated>2007-05-19T22:35:37Z</updated>

		<summary type="html">&lt;p&gt;Aris: /* Contact */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__TOC__&lt;br /&gt;
==Process==&lt;br /&gt;
The software development process in MAGNet typically takes place in 2 phases:&lt;br /&gt;
* Prototype tools are developed and field-tested at various investigator labs in the context of Core 1, 2 and 3 activities. These tools are usually command line versions, coded in an assortment of languages (C, C++, Perl, etc) and/or computational environments (MATLAB, R). In the majority of cases, analysis results are represented as text files that are either inspected with text editors or are transformed (through custom programs) to properly formatted inputs for downstream analysis/visualization tools.&lt;br /&gt;
* When prototypes have reached a satisfactory level of maturity the software engineering group takes over the task of integrating them into the Center's bioinformatics platform, geWorkbench, http://www.geworkbench.org. Operationally, each integration activity is carried out following a number of steps deriving from a standard software life cycle process:&lt;br /&gt;
:# '''Collection of business requirements''': an initial meeting is arranged with the investigator whose prototype is slated for integration. During the meeting high level requirements are collected, outlining the intended functionality of the integrated component. Such a meeting usually involves (i) a demo of geWorkbench, (ii) an overview presentation outlining the science behind the investigator's prototype, (iii) a demo of the prototype, and (iv) a brainstorming session aimed at identifying the most meaningful manner to integrate the prototype into geWorkbench so that the prototype leverages as much as possible of the functionality already present in geWorkbench.&lt;br /&gt;
:# '''Development of Use case document''':  based on the information collected at the investigator meeting a detailed Use Case requirements document is developed which describes precisely what the user interface will look like and how users will interact with it. The Use Case document is subject to review and approval by the investigator. An example of such a document can be found [[Media:CNKB_UC.pdf | here]].&lt;br /&gt;
:# '''Development of user interface/computational service''': based on the requirements laid out in the use case document, the development team (i) develops the user interface for the integrated version of the investigator tool, and (ii) ports any associated compute server to the grid-enabled server framework used by geWorkbench (which is based on the caGrid software, https://cabig.nci.nih.gov/workspaces/Architecture/caGrid, a grid middleware layer developed in the context of the caBIG initiative). Unit tests are an integral part of the development process and are run automatically on a nightly basis.&lt;br /&gt;
:# '''System Testing''': After the prototype in integrated, an integration and system testing cycle is performed. Our group uses a formal testing process based on detailed test scripts whose execution status is tracked in a custom database we have developed for this purpose. In this database, failed scripts are linked to their associated defect reports within our defect management system, http://wiki.c2b2.columbia.edu/mantis/.&lt;br /&gt;
:# '''Documentation''': The final step in the development process is the generation of detailed end-user documentation, in the form of on-line help, tutorials, end-user guides and training slides. Links to these materials are available at: http://wiki.c2b2.columbia.edu/workbench/index.php/Project_Documentation. &lt;br /&gt;
&lt;br /&gt;
geWorkbench is an open-source platform and we encourage and welcome code contributions by the community. To facilitate such contributions we have made the entire code base freely available to everyone through our community development project site, http://gforge.nci.nih.gov/projects/geworkbench/, registered with the NCI GForge project. This site enables us to leverage the infrastructure support that GForge offers to collaborative development projects, including access to a CVS server, streamlined setup of mailing lists and user-forums, posting software releases, etc.&lt;br /&gt;
&lt;br /&gt;
==License==&lt;br /&gt;
At present, geWorkbench in made available under the licensing terms stated at http://wiki.c2b2.columbia.edu/workbench/index.php/GeWorkbench_License. We are currently in the process of modularizing our licensing , in order to accommodate the needs of different labs/tools.&lt;br /&gt;
&lt;br /&gt;
==Contact==&lt;br /&gt;
The software engineering group at MAGNet is managed by Aris Floratos (floratos at c2b2 dot columbia dot edu.&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10628</id>
		<title>SDIWG:Software Engineering At MAGNet</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10628"/>
		<updated>2007-05-19T22:30:22Z</updated>

		<summary type="html">&lt;p&gt;Aris: /* Process */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__TOC__&lt;br /&gt;
==Process==&lt;br /&gt;
The software development process in MAGNet typically takes place in 2 phases:&lt;br /&gt;
* Prototype tools are developed and field-tested at various investigator labs in the context of Core 1, 2 and 3 activities. These tools are usually command line versions, coded in an assortment of languages (C, C++, Perl, etc) and/or computational environments (MATLAB, R). In the majority of cases, analysis results are represented as text files that are either inspected with text editors or are transformed (through custom programs) to properly formatted inputs for downstream analysis/visualization tools.&lt;br /&gt;
* When prototypes have reached a satisfactory level of maturity the software engineering group takes over the task of integrating them into the Center's bioinformatics platform, geWorkbench, http://www.geworkbench.org. Operationally, each integration activity is carried out following a number of steps deriving from a standard software life cycle process:&lt;br /&gt;
:# '''Collection of business requirements''': an initial meeting is arranged with the investigator whose prototype is slated for integration. During the meeting high level requirements are collected, outlining the intended functionality of the integrated component. Such a meeting usually involves (i) a demo of geWorkbench, (ii) an overview presentation outlining the science behind the investigator's prototype, (iii) a demo of the prototype, and (iv) a brainstorming session aimed at identifying the most meaningful manner to integrate the prototype into geWorkbench so that the prototype leverages as much as possible of the functionality already present in geWorkbench.&lt;br /&gt;
:# '''Development of Use case document''':  based on the information collected at the investigator meeting a detailed Use Case requirements document is developed which describes precisely what the user interface will look like and how users will interact with it. The Use Case document is subject to review and approval by the investigator. An example of such a document can be found [[Media:CNKB_UC.pdf | here]].&lt;br /&gt;
:# '''Development of user interface/computational service''': based on the requirements laid out in the use case document, the development team (i) develops the user interface for the integrated version of the investigator tool, and (ii) ports any associated compute server to the grid-enabled server framework used by geWorkbench (which is based on the caGrid software, https://cabig.nci.nih.gov/workspaces/Architecture/caGrid, a grid middleware layer developed in the context of the caBIG initiative). Unit tests are an integral part of the development process and are run automatically on a nightly basis.&lt;br /&gt;
:# '''System Testing''': After the prototype in integrated, an integration and system testing cycle is performed. Our group uses a formal testing process based on detailed test scripts whose execution status is tracked in a custom database we have developed for this purpose. In this database, failed scripts are linked to their associated defect reports within our defect management system, http://wiki.c2b2.columbia.edu/mantis/.&lt;br /&gt;
:# '''Documentation''': The final step in the development process is the generation of detailed end-user documentation, in the form of on-line help, tutorials, end-user guides and training slides. Links to these materials are available at: http://wiki.c2b2.columbia.edu/workbench/index.php/Project_Documentation. &lt;br /&gt;
&lt;br /&gt;
geWorkbench is an open-source platform and we encourage and welcome code contributions by the community. To facilitate such contributions we have made the entire code base freely available to everyone through our community development project site, http://gforge.nci.nih.gov/projects/geworkbench/, registered with the NCI GForge project. This site enables us to leverage the infrastructure support that GForge offers to collaborative development projects, including access to a CVS server, streamlined setup of mailing lists and user-forums, posting software releases, etc.&lt;br /&gt;
&lt;br /&gt;
==License==&lt;br /&gt;
At present, geWorkbench in made available under the licensing terms stated at http://wiki.c2b2.columbia.edu/workbench/index.php/GeWorkbench_License. We are currently in the process of modularizing our licensing , in order to accommodate the needs of different labs/tools.&lt;br /&gt;
&lt;br /&gt;
==Contact==&lt;br /&gt;
The software engineering group at MAGNet is managed by [mailto:%66%6C%6F%72%61%74%6F%73%40%63%32%62%32%2E%63%6F%6C%75%6D%62%69%61%2E%65%64%75 Aris Floratos].&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10627</id>
		<title>SDIWG:Software Engineering At MAGNet</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10627"/>
		<updated>2007-05-19T22:26:39Z</updated>

		<summary type="html">&lt;p&gt;Aris: /* Process */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__TOC__&lt;br /&gt;
==Process==&lt;br /&gt;
The software development process in MAGNet typically takes place in 2 phases:&lt;br /&gt;
* Prototype tools are developed and field-tested at various investigator labs in the context of Core 1, 2 and 3 activities. These tools are usually command line versions, coded in an assortment of languages (C, C++, Perl, etc) and/or computational environments (MATLAB, R). In the majority of cases, analysis results are represented as text files that are either inspected with text editors or are transformed (through custom programs) to properly formatted inputs for downstream analysis/visualization tools.&lt;br /&gt;
* When prototypes have reached a satisfactory level of maturity the software engineering group takes over the task of integrating them into the Center's bioinformatics platform, geWorkbench, http://www.geworkbench.org. Operationally, each integration activity is carried out following a number of steps deriving from a standard software life cycle process:&lt;br /&gt;
:# '''Collection of business requirements''': an initial meeting is arranged with the investigator whose prototype is slated for integration. During the meeting high level requirements are collected, outlining the intended functionality of the integrated component. Such a meeting usually involves (i) a demo of geWorkbench, (ii) an overview presentation outlining the science behind the investigator's prototype, (iii) a demo of the prototype, and (iv) a brainstorming session aimed at identifying the most meaningful manner to integrate the prototype into geWorkbench so that the prototype leverages as much as possible of the functionality already present in geWorkbench.&lt;br /&gt;
:# '''Development of Use case document''':  based on the information collected at the investigator meeting a detailed Use Case requirements document is developed which describes precisely what the user interface will look like and how users will interact with it. The Use Case document is subject to review and approval by the investigator. An example of such a document can be found [[Media:CNKB_UC.pdf | here]].&lt;br /&gt;
:# '''Development of user interface/computational service''': based on the requirements laid out in the use case document, the development team (a) develops the user interface for the integrated version of the investigator tool, and (2) ports any associated computate server to the grid-enabled server framework used by geWorkbench (which is based on the caGrid software, https://cabig.nci.nih.gov/workspaces/Architecture/caGrid, a grid middleware layer developed by the caBIG initiative). Unit tests are an integral part of the development process and are run automatically on a nightly basis.&lt;br /&gt;
:# '''System Testing''': After the prototype in integrated a formal integration and system testing cycle ensues. Our group uses a formal testing process based on detailed test scripts whose execution status is tracked in a custom database we have developed for this purpose. In this database, failed scripts are linked to their associated defect reports within our defect management system, http://wiki.c2b2.columbia.edu/mantis/.&lt;br /&gt;
:# '''Documentation''': The final step in the development process is the generation of detailed end-user documentation, in the form of on-line help, tutorials, end-user guides and training slides. Links to these materials are available at: http://wiki.c2b2.columbia.edu/workbench/index.php/Project_Documentation. &lt;br /&gt;
&lt;br /&gt;
geWorkbench is an open-source platform and we encourage and welcome code contributions by the community. To facilitate such contributions we have made the entire code base freely available to everyone through our community development project site, http://gforge.nci.nih.gov/projects/geworkbench/, registered with the NCI GForge project. This site enables us to leverage the infrastructure support that GForge offers to collaborative development projects, including access to a CVS server, streamlined setup of mailing lists and user-forums, posting software releases, etc.&lt;br /&gt;
&lt;br /&gt;
==License==&lt;br /&gt;
At present, geWorkbench in made available under the licensing terms stated at http://wiki.c2b2.columbia.edu/workbench/index.php/GeWorkbench_License. We are currently in the process of modularizing our licensing , in order to accommodate the needs of different labs/tools.&lt;br /&gt;
&lt;br /&gt;
==Contact==&lt;br /&gt;
The software engineering group at MAGNet is managed by [mailto:%66%6C%6F%72%61%74%6F%73%40%63%32%62%32%2E%63%6F%6C%75%6D%62%69%61%2E%65%64%75 Aris Floratos].&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=File:CNKB_UC.pdf&amp;diff=10626</id>
		<title>File:CNKB UC.pdf</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=File:CNKB_UC.pdf&amp;diff=10626"/>
		<updated>2007-05-19T22:24:30Z</updated>

		<summary type="html">&lt;p&gt;Aris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10625</id>
		<title>SDIWG:Software Engineering At MAGNet</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10625"/>
		<updated>2007-05-19T22:21:17Z</updated>

		<summary type="html">&lt;p&gt;Aris: /* Process */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__TOC__&lt;br /&gt;
==Process==&lt;br /&gt;
The software development process in MAGNet typically takes place in 2 phases:&lt;br /&gt;
* Prototype tools are developed and field-tested at various investigator labs in the context of Core 1, 2 and 3 activities. These tools are usually command line versions, coded in an assortment of languages (C, C++, Perl, etc) and/or computational environments (MATLAB, R). In the majority of cases, analysis results are represented as text files that are either inspected with text editors or are transformed (through custom programs) to properly formatted inputs for downstream analysis/visualization tools.&lt;br /&gt;
* When prototypes have reached a satisfactory level of maturity the software engineering group takes over the task of integrating them into the Center's bioinformatics platform, geWorkbench, http://www.geworkbench.org. Operationally, each integration activity is carried out following a number of steps deriving from a standard software life cycle process:&lt;br /&gt;
:# '''Collection of business requirements''': an initial meeting is arranged with the investigator whose prototype is slated for integration. During the meeting high level requirements are collected, outlining the intended functionality of the integrated component. Such a meeting usually involves (i) a demo of geWorkbench, (ii) an overview presentation outlining the science behind the investigator's prototype, (iii) a demo of the prototype, and (iv) a brainstorming session aimed at identifying the most meaningful manner to integrate the prototype into geWorkbench so that the prototype leverages as much as possible of the functionality already present in geWorkbench.&lt;br /&gt;
:# '''Development of Use case document''':  based on the information collected at the investigator meeting a detailed Use Case requirements document is developed which describes precisely what the user interface will look like and how users will interact with it. The Use Case document is subject to review and approval by the investigator. An example of such a document can be found at http://magnet.c2b2.columbia.edu/AnnualReport/Y2/CNKB_UC.pdf.&lt;br /&gt;
:# '''Development of user interface/computational service''': based on the requirements laid out in the use case document, the development team (a) develops the user interface for the integrated version of the investigator tool, and (2) ports any associated computate server to the grid-enabled server framework used by geWorkbench (which is based on the caGrid software, https://cabig.nci.nih.gov/workspaces/Architecture/caGrid, a grid middleware layer developed by the caBIG initiative). Unit tests are an integral part of the development process and are run automatically on a nightly basis.&lt;br /&gt;
:# '''System Testing''': After the prototype in integrated a formal integration and system testing cycle ensues. Our group uses a formal testing process based on detailed test scripts whose execution status is tracked in a custom database we have developed for this purpose. In this database, failed scripts are linked to their associated defect reports within our defect management system, http://wiki.c2b2.columbia.edu/mantis/.&lt;br /&gt;
:# '''Documentation''': The final step in the development process is the generation of detailed end-user documentation, in the form of on-line help, tutorials, end-user guides and training slides. Links to these materials are available at: http://wiki.c2b2.columbia.edu/workbench/index.php/Project_Documentation. &lt;br /&gt;
&lt;br /&gt;
geWorkbench is an open-source platform and we encourage and welcome code contributions by the community. To facilitate such contributions we have made the entire code base freely available to everyone through our community development project site, http://gforge.nci.nih.gov/projects/geworkbench/, registered with the NCI GForge project. This site enables us to leverage the infrastructure support that GForge offers to collaborative development projects, including access to a CVS server, streamlined setup of mailing lists and user-forums, posting software releases, etc.&lt;br /&gt;
&lt;br /&gt;
==License==&lt;br /&gt;
At present, geWorkbench in made available under the licensing terms stated at http://wiki.c2b2.columbia.edu/workbench/index.php/GeWorkbench_License. We are currently in the process of modularizing our licensing , in order to accommodate the needs of different labs/tools.&lt;br /&gt;
&lt;br /&gt;
==Contact==&lt;br /&gt;
The software engineering group at MAGNet is managed by [mailto:%66%6C%6F%72%61%74%6F%73%40%63%32%62%32%2E%63%6F%6C%75%6D%62%69%61%2E%65%64%75 Aris Floratos].&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10613</id>
		<title>SDIWG:Software Engineering At MAGNet</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10613"/>
		<updated>2007-05-18T22:38:07Z</updated>

		<summary type="html">&lt;p&gt;Aris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__TOC__&lt;br /&gt;
==Process==&lt;br /&gt;
The software development process in MAGNet typically takes place in 2 phases:&lt;br /&gt;
* Prototype tools are developed and field-tested at various investigator labs in the context of Core 1, 2 and 3 activities. These tools are usually command line versions, coded in an assortment of languages (C, C++, Perl, etc) and/or computational environments (MATLAB, R). In the majority of cases, analysis results are represented as text files that are either inspected with text editors or are transformed (through custom programs) to properly formatted inputs for downstream analysis/visualization tools.&lt;br /&gt;
* When prototypes have reached a satisfactory level of maturity the software engineering group takes over the task of integrating them into the Center's bioinformatics platform, geWorkbench, http://www.geworkbench.org. Operationally, each integration activity is carried out following a number of steps deriving from a standard software life cycle process:&lt;br /&gt;
## '''Collection of business requirements''': an initial meeting is arranged with the investigator whose prototype is slated for integration. During the meeting high level requirements are collected, outlining the intended functionality of the integrated component. Such a meeting usually involves (i) a demo of geWorkbench, (ii) an overview presentation outlining the science behind the investigator's prototype, (iii) a demo of the prototype, and (iv) a brainstorming session aimed at identifying the most meaningful manner to integrate the prototype into geWorkbench so that the prototype leverages as much as possible of the functionality already present in geWorkbench.&lt;br /&gt;
## '''Development of Use case document''':  based on the information collected at the investigator meeting a detailed Use Case requirements document is developed which describes precisely what the user interface will look like and how users will interact with it. The Use Case document is subject to review and approval by the investigator. An example of such a document can be found at http://magnet.c2b2.columbia.edu/AnnualReport/Y2/CNKB_UC.pdf.&lt;br /&gt;
## '''Development of user interface/computational service''': based on the requirements laid out in the use case document, the development team (a) develops the user interface for the integrated version of the investigator tool, and (2) ports any associated computate server to the grid-enabled server framework used by geWorkbench (which is based on the caGrid software, https://cabig.nci.nih.gov/workspaces/Architecture/caGrid, a grid middleware layer developed by the caBIG initiative). Unit tests are an integral part of the development process and are run automatically on a nightly basis.&lt;br /&gt;
## '''System Testing''': After the prototype in integrated a formal integration and system testing cycle ensues. Our group uses a formal testing process based on detailed test scripts whose execution status is tracked in a custom database we have developed for this purpose. In this database, failed scripts are linked to their associated defect reports within our defect management system, http://wiki.c2b2.columbia.edu/mantis/.&lt;br /&gt;
## '''Documentation''': The final step in the development process is the generation of detailed end-user documentation, in the form of on-line help, tutorials, end-user guides and training slides. Links to these materials are available at: http://wiki.c2b2.columbia.edu/workbench/index.php/Project_Documentation. &lt;br /&gt;
&lt;br /&gt;
geWorkbench is an open-source platform and we encourage and welcome code contributions by the community. To facilitate such contributions we have made the entire code base freely available to everyone through our community development project site, http://gforge.nci.nih.gov/projects/geworkbench/, registered with the NCI GForge project. This site enables us to leverage the infrastructure support that GForge offers to collaborative development projects, including access to a CVS server, streamlined setup of mailing lists and user-forums, posting software releases, etc.&lt;br /&gt;
&lt;br /&gt;
==License==&lt;br /&gt;
At present, geWorkbench in made available under the licensing terms stated at http://wiki.c2b2.columbia.edu/workbench/index.php/GeWorkbench_License. We are currently in the process of modularizing our licensing , in order to accommodate the needs of different labs/tools.&lt;br /&gt;
&lt;br /&gt;
==Contact==&lt;br /&gt;
The software engineering group at MAGNet is managed by [mailto:%66%6C%6F%72%61%74%6F%73%40%63%32%62%32%2E%63%6F%6C%75%6D%62%69%61%2E%65%64%75 Aris Floratos].&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10612</id>
		<title>SDIWG:Software Engineering At MAGNet</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:Software_Engineering_At_MAGNet&amp;diff=10612"/>
		<updated>2007-05-18T22:32:42Z</updated>

		<summary type="html">&lt;p&gt;Aris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__TOC__&lt;br /&gt;
==Process==&lt;br /&gt;
The software development process in MAGNet typically takes place in 2 phases:&lt;br /&gt;
* Prototype tools are developed and field-tested at various investigator labs in the context of Core 1, 2 and 3 activities. These tools are usually command line versions, coded in an assortment of languages (C, C++, Perl, etc) and/or computational environments (MATLAB, R). In the majority of cases, analysis results are represented as text files that are either inspected with text editors or are transformed (through custom programs) to properly formatted inputs for downstream analysis/visualization tools.&lt;br /&gt;
* When prototypes have reached a satisfactory level of maturity the software engineering group takes over the task of integrating them into the Center's bioinformatics platform, geWorkbench, http://www.geworkbench.org. Operationally, each integration activity is carried out following a number of steps deriving from a standard software life cycle process:&lt;br /&gt;
## '''Collection of business requirements''': an initial meeting is arranged with the investigator whose prototype is slated for integration. During the meeting high level requirements are collected, outlining the intended functionality of the integrated component. Such a meeting usually involves (i) a demo of geWorkbench, (ii) an overview presentation outlining the science behind the investigator's prototype, (iii) a demo of the prototype, and (iv) a brainstorming session aimed at identifying the most meaningful manner to integrate the prototype into geWorkbench so that the prototype leverages as much as possible of the functionality already present in geWorkbench.&lt;br /&gt;
## '''Development of Use case document''':  based on the information collected at the investigator meeting a detailed Use Case requirements document is developed which describes precisely what the user interface will look like and how users will interact with it. The Use Case document is subject to review and approval by the investigator. An example of such a document can be found at http://magnet.c2b2.columbia.edu/AnnualReport/Y2/CNKB_UC.pdf.&lt;br /&gt;
## '''Development of user interface/computational service''': based on the requirements laid out in the use case document, the development team (a) develops the user interface for the integrated version of the investigator tool, and (2) ports any associated computate server to the grid-enabled server framework used by geWorkbench (which is based on the caGrid software, https://cabig.nci.nih.gov/workspaces/Architecture/caGrid, a grid middleware layer developed by the caBIG initiative). Unit tests are an integral part of the development process and are run automatically on a nightly basis.&lt;br /&gt;
## '''System Testing''': After the prototype in integrated a formal integration and system testing cycle ensues. Our group uses a formal testing process based on detailed test scripts whose execution status is tracked in a custom database we have developed for this purpose. In this database, failed scripts are linked to their associated defect reports within our defect management system, http://wiki.c2b2.columbia.edu/mantis/.&lt;br /&gt;
## '''Documentation''': The final step in the development process is the generation of detailed end-user documentation, in the form of on-line help, tutorials, end-user guides and training slides. Links to these materials are available at: http://wiki.c2b2.columbia.edu/workbench/index.php/Project_Documentation. &lt;br /&gt;
&lt;br /&gt;
geWorkbench is an open-source platform and we encourage and welcome code contributions by the community. To facilitate such contributions we have made the entire code base freely available to everyone through our community development project site, http://gforge.nci.nih.gov/projects/geworkbench/, registered with the NCI GForge project. This site enables us to leverage the infrastructure support that GForge offers to collaborative development projects, including access to a CVS server, streamlined setup of mailing lists and user-forums, posting software releases, etc.&lt;br /&gt;
&lt;br /&gt;
==Contact==&lt;br /&gt;
The software engineering group at MAGNet is managed by [mailto:%66%6C%6F%72%61%74%6F%73%40%63%32%62%32%2E%63%6F%6C%75%6D%62%69%61%2E%65%64%75 Aris Floratos].&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10358</id>
		<title>SDIWG:NCBC Software Classification MAGNet Examples</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10358"/>
		<updated>2007-05-15T04:52:34Z</updated>

		<summary type="html">&lt;p&gt;Aris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;===DelPhi===&lt;br /&gt;
* '''Description''': DelPhi provides numerical solutions to the Poisson-Boltzmann equation (both linear and nonlinear form) for molecules of arbitrary shape and charge distribution. The current version is fast (the best relaxation parameter is estimated at run time), accurate (calculation of the electrostatic free energy is less dependent on the resolution of the lattice) and can handle extremely high lattice dimensions. It also includes flexible features for assigning different dielectric constants to different regions of space and treating systems containing mixed salt solutions.&lt;br /&gt;
* '''Data Input''': DelPhi takes as input a coordinate file format of a molecule or equivalent data for geometrical objects and/or charge distributions&lt;br /&gt;
* '''Data Output''': electrostatic potential in and around the system&lt;br /&gt;
* '''Implementation Language''': Fortran and C&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''':  E.Alexov,  R.Fine, M.K.Gilson, A.Nicholls, W.Rocchia, K.Sharp, and B. Honig.&lt;br /&gt;
* '''Platforms Tested''': Unix-SGI IRIX, linux, PC (requires Fortran and C compilers), AIX IBM version and Mac.&lt;br /&gt;
* '''License''': Freely available to academia; pay model for commercial users. &lt;br /&gt;
* '''Keywords''': Finite Difference Poisson-Boltzman Solver&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/delphi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Numerical Calculation of Electrostatic Potential&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP===&lt;br /&gt;
* '''Description''': A molecular visualization and analysis program. It is particularly useful for the display and manipulation of the surfaces of molecules and their electrostatic properties.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi&lt;br /&gt;
* '''Data Output''': molecular graphics.&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': v1.3.6 .Stable public release.&lt;br /&gt;
* '''Authors''': Anthony Nicholls and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': SGI machines: irix 5.x and 6.x (INDYs, INDIGOs including Impact, Octane and O2) systems.&lt;br /&gt;
* '''License''':  Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
===Nest===&lt;br /&gt;
* '''Description''': Modeling protein structure based on a sequence-template alignment. The current server works only for modeling with a single template. Part of jackal, which can be downloaded.&lt;br /&gt;
* '''Data Input''': pir and PDB files&lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Xiang, Z. and Honig, B.&lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': modeling, protein structure, sequence-template alignment.&lt;br /&gt;
* '''URL''': http://honiglab.cpmc.columbia.edu/cgi-bin/jackal/nest.cgi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===JACKAL===&lt;br /&gt;
* '''Description''': Jackal is a collection of programs designed for the modeling and analysis of protein structures. Its core program is a versatile homology modeling package nest. JACKAL has the following capabilities: 1) comparative modeling based on single, composite or multiple templates; 2) side-chain prediction; 3) modeling residue mutation, insertion or deletion; 4) loop prediction; 5) structure refinement; 6) reconstruction of protein missing atoms;7) reconstruction of protein missing residues; 8) prediction of hydrogen atoms; 9) fast calculation of solvent accessible surface area; 10) structure superimposition.       &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Version: 1.5 as of Oct, 20, 2002, Stable public release.&lt;br /&gt;
* '''Authors''': Z. Xiang and B. Honig&lt;br /&gt;
* '''Platforms Tested''': SGI 6.5, Intel Linux and Sun solaris&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': Protein Structure Modeling&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/jackal&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP2===&lt;br /&gt;
* '''Description''': GRASP2 is an updated version of the GRASP program used for macromolecular structure and surface visualization, contains a large number of new features and scientific tools: Enhanced GUI; Structure alignment and domain database scanning; A gaussian surface generator and new surface coloring schemes; Sequence visualization and alignment; Completed work can be stored in &amp;quot;project files;  Among the many objects that can be stored in a project file are views of the structure; defined subsets, surfaces; Direct printing to printers at full printer resolution.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi, sequence alignments.&lt;br /&gt;
* '''Data Output''': molecular graphics, structural alignments.&lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Donald Petrey and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp2&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===PrISM===&lt;br /&gt;
* '''Description''': PrISM is an integrated computational system where computational tools are implemented for protein sequence and structure analysis and modeling.&lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Wang, L, Yang, A. S. &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''':SGI-irix, Intel-linux &lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': protein analysis/modeling &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/PrISM/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Protein-DNA interface alignment===&lt;br /&gt;
* '''Description''': The protein-DNA alignment software allows one to align the interfacial amino acids from two protein-DNA complexes based on the geometric relationship of each amino acid to its local DNA. &lt;br /&gt;
* '''Data Input''': two PDB files that both contain protein-DNA complexes&lt;br /&gt;
* '''Data Output''': The programs will output the aligned residues and their corresponding residue-residue similarity scores, s(i,j).&lt;br /&gt;
* '''Implementation Language''': C++ and Perl&lt;br /&gt;
* '''Version, Date, Stage''':Stable public release. &lt;br /&gt;
* '''Authors''': Siggers, T.W., Silkov, A &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''': Linux&lt;br /&gt;
* '''License''': Freely available to academia &lt;br /&gt;
* '''Keywords''': protein-DNA interface &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/intfc_aln&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
===SURFace===&lt;br /&gt;
* '''Description''': SURFace algorithms are programs that calculate solvent accessible surface area and curvature corrected solvent accessible surface area &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': &lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Nicholls, A., Sharp, K., Sridharan, S. and Honig, B.  &lt;br /&gt;
* '''Platforms Tested''': SGI&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': solvent accessible surface area  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/surf/&lt;br /&gt;
* '''Organization''': MAGNet &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Caculation of Solvent Accessible Area&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Target Explorer===&lt;br /&gt;
* '''Description''': Automated process of prediction of complex regulatory elements for specified set of transcription factors in Drosophila melanogaster genome. Target Explorer is a complex tool with user-friendly self-explanatory Web-interface that allows to user:  1. create customized library of TF binding site matrices based on user defined sets of training sequences; 2. search for new clusters of binding sites for specified set of TFs; 3.extract annotation for potential target genes.    &lt;br /&gt;
* '''Data Input''': genomic sequences&lt;br /&gt;
* '''Data Output''': clusters of known binding sites &lt;br /&gt;
* '''Implementation Language''': perl, cgi&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release. &lt;br /&gt;
* '''Authors''': Sosinsky A, Bonin CP, Mann RS, Honig B.   &lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': prediction of binding sites for transcription factors  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/Target_Explorer/&lt;br /&gt;
* '''Organization''': MAGNet  &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis--&amp;gt; Sequence Annotation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===MEDUSA and Gorgon===&lt;br /&gt;
* '''Description''': MEDUSA is an algorithm for learning predictive models of transcriptional gene regulation from gene expression and promoter sequence data.  By using a statistical learning approach based on boosting, MEDUSA learns cis regulatory motifs, condition-specific regulators, and regulatory programs that predict the differential expression of target genes.  The regulatory program is specified as an alternating decision tree (ADT).  The Java implementation of MEDUSA allows a number of visualizations of the regulatory program and other inferred regulatory information, implemented in the accompanying Gorgon tool, including hits of significant and condition-specific motifs along the promoter sequences of target genes and regulatory network figures viewable in Cytoscape.      &lt;br /&gt;
* '''Data Input''': Discretized (up/down/baseline) gene expression data in plain text format, promoter sequences in FASTA format, list of candidate transcriptional regulators and signal transducers in plain text format.&lt;br /&gt;
* '''Data Output''': Regulatory program represented as a Java serialized object file readable by Gorgon and as a human readable XML file.  Gorgon currently generates views of learned PSSMs, positional hits along promoter sequences, and views of the ADT as HTML files, and generates network figures as Cytoscape format files.&lt;br /&gt;
* '''Implementation Language''': Java (prototyped in MATLAB) &lt;br /&gt;
* '''Version, Date, Stage''': Version 2.0, July 2006, pre-release beta version; Version 1.0 (MATLAB), April 2005, stable public release  &lt;br /&gt;
* '''Authors''': David Quigley, Manuel Middendorf, Steve Lianoglou, Anshul Kundaje, Yoav Freund, Chris Wiggins, Christina Leslie   &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux, Mac OS X&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/medusa (MATLAB),http://compbio.sytes.net:8090/medusa (Java beta version)&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;br /&gt;
&lt;br /&gt;
===String kernel package===&lt;br /&gt;
* '''Description''':The string kernel package contains implementations for the mismatch and profile string kernels for use with support vector machine (SVM) classifiers for protein sequence classification.  Both kernels compute similarity between protein sequences based on common occurrences of k-length subsequences (&amp;quot;k-mers&amp;quot;) counted with substitutions.  Kernel functions for protein sequence data enable the training of SVMs for a range of prediction problems, in particular protein structural class prediction and remote homology detection.  A version of the Spider MATLAB machine learning package is also bundled with the code, which allows users to train SVMs and evaluate performance on test sets with the packaged software.&lt;br /&gt;
* '''Data Input''': The mismatch kernel requires sequence data in FASTA format. The profile string kernel uses probabilistic profiles, such as those produced by PSI-BLAST, in place of the original sequences.  The Spider SVM implementation requires both the kernel matrix and a label file of binary or multi-class labels for the training data; this data must be loaded into MATLAB variables before using Spider routing.&lt;br /&gt;
* '''Data Output''':The kernel code produces a kernel matrix for the input data in tab-delimited text format.  The Spider package trains SVMs and stores the learns classifier and results from applying the classifier on test data as MATLAB objects.&lt;br /&gt;
* '''Implementation Language''': String kernel code is implemented in C. Spider is a set of object-oriented MATLAB routines. &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.2, September 2004, stable public release &lt;br /&gt;
* '''Authors''': Eleazar Eskin, Rui Kuang, Eugene Ie, Ke Wang, Jason Weston, Bill Noble, Christina Leslie  &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/string-kernels&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification &lt;br /&gt;
&lt;br /&gt;
===MatrixREDUCE===&lt;br /&gt;
* '''Description''': Regulation of gene expression by a transcription factor requires physical interaction between the factor and the DNA, which can be described by astatistical mechanical model. Based on this model, the MatrixREDUCE algorithm uses genome-wide occupancy data for a transcription factor (e.g.ChIP-chip or mRNA expression data) and associated nucleotide sequences to discover the sequence-specific binding affinity of the transcription factor. The sequence specificity of the transcription factor's DNA-binding domain is modeled using a position-specific affinity matrix (PSAM), representing the change in the binding affinity (Kd) whenever a specific position within a reference binding sequence is mutated. The PSAM can be transformed into affinity logo for visualization using the utility program AffinityLogo, and a MatrixREDUCE run can be summarized in an easy-to-navigate webpage using HTMLSummary. &lt;br /&gt;
* '''Data Input''': sequence file in FASTA format; and expression data file in tab-delimited text format.&lt;br /&gt;
* '''Data Output''': PSAMs in numeric and graphical format, parameters of the fitted model, and an HTML summary page.&lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of Numerical Recipes routines.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.0, July 10, 2006, extensively tested in lab.  &lt;br /&gt;
* '''Authors''': Barrett Foat, Xiang-Jun Lu, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': position-specific affinity matrix, binding affinity, cis-regulatory element, expression data, ChIP-chip, transcription factor &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/MatrixREDUCE&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===T-profiler===&lt;br /&gt;
* '''Description''': T-profiler is a web-based tool that uses the t-test to score changes in the average activity of pre-defined groups of genes. The gene groups are defined based on Gene Ontology categorization, ChIP-chip experiments, upstream matches to a consensus transcription factor binding motif, and location on the same chromosome, respectively. If desired, an iterative procedure can be used to select a single, optimal representative from sets of overlapping gene groups. A jack-knife procedure is used to make calculations more robust against outliers. T-profiler makes it possible to interpret microarray data in a way that is both intuitive and statistically rigorous, without the need to combine experiments or choose parameters.   &lt;br /&gt;
* '''Data Input''': Currently, gene expression data from Saccharomyces cerevisiae and Candida albicans are supported. &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': T-profiler is written in PHP, data is managed by a MYSQL database server  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': André Boorsma, Barrett C. Foat, Daniel Vis, Frans Klis, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Web-based application&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': gene expression, transcriptome,  ChIP-chip,  Gene Ontology &lt;br /&gt;
* '''URL''': http://www.t-profiler.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Network characterization &lt;br /&gt;
&lt;br /&gt;
===TranscriptionDetector===&lt;br /&gt;
* '''Description''':A tool for finding probes measuring significantly expressed loci in a genomic array experiment. Given expression data from some tiling array experiment, TranscriptionDetector decides the likelihood that a probe is detecting transcription from the locus in which it resides. Probabilities are assigned by making use of a background signal intensity distribution from a set of negative control probes. This tool is useful for the functional annotation of genomes as it allows for the discovery of novel transcriptional units independently of any genomic annotation.    &lt;br /&gt;
* '''Data Input''': Expression data (GEO or other platforms) and designation of which probes represent negative controls and which are data probes. &lt;br /&gt;
* '''Data Output''': A text file with a list of probes  corresponding to significantly expressed loci. &lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of GSL.  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': Xiang-Jun Lu, Gabor Halasz, Marinus F. van Batenburg  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': tiling arrays, expression, transcriptome &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/TranscriptionDetector/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===PhenoGO===&lt;br /&gt;
* '''Description''': PhenoGO adds phenotypic contextual information to existing associations between gene products and Gene Ontology (GO) terms as specified in GO Annotations (GOA). PhenoGO utilizes an existing Natural Language Processing (NLP) system, called BioMedLEE, an existing knowledge-based phenotype organizer system (PhenOS) in conjunction with MeSH indexing and established biomedical ontologies. The system also encodes the context to identifiers that are associated in different biomedical ontologies, including the UMLS, Cell Ontology, Mouse Anatomy, NCBI taxonomy, GO, and Mammalian Phenotype Ontology. In addition, PhenoGO was evaluated for coding of anatomical and cellular information and assigning the coded phenotypes to the correct GOA; results obtained show that PhenoGO has a precision of 91% and recall of 92%, demonstrating that the PhenoGO NLP system can accurately encode a large number of anatomical and cellular ontologies to GO annotations. The PhenoGO Database may be accessed at www.phenogo.org.    &lt;br /&gt;
* '''Data Input''': Gene Ontology Annotations Files and Medline Abstracts  &lt;br /&gt;
* '''Data Output''': XML file and www.phenogo.org Web Portal&lt;br /&gt;
* '''Implementation Language''': A variety of modules, the web portal is in Java and MySQL, the computational terminology component (phenOS) is written in Perl scripts that queries tables in IBM DB2,  the natural language processing component is written in PROLOG.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 2, Feb 2006  &lt;br /&gt;
* '''Authors''': Yves Lussier and Carol Friedman are the principal investigators. The programmers are Jianrong Li, Lee Sam, and Tara Borlawsky   &lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': Phenotypic integration, computational phenotypes &lt;br /&gt;
* '''URL''': http://www.phenogo.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Biotool --&amp;gt; Data Management --&amp;gt; Information retrieval, traversal and querying; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Natural Language Processing &lt;br /&gt;
&lt;br /&gt;
===MINDY===&lt;br /&gt;
* '''Description''': Given a transcription factor of interest, MINDY uses a large set of gene expression profile data to identify potential post-transcriptional modulators of the transcription factor's activity. MINDY is based on a three-way statistical interaction model that captures the post-transcriptional regulatory event where the ability of a transcription factor to activate/repress its target genes is monotonically controlled by a potential modulator gene. &lt;br /&gt;
* '''Data Input''': Gene expression data in the EXP format, and a user-specified transcription factor of interest&lt;br /&gt;
* '''Data Output''': Lists of the putative modulators and target genes of the transcription factor, and the modulatory interactions involving them&lt;br /&gt;
* '''Implementation Language''': C++ and MATLAB, Java&lt;br /&gt;
* '''Version, Date, Stage''': Stable release, April 2007  &lt;br /&gt;
* '''Authors''': Kai Wang, Ilya Nemenman, Adam Margolin, Riccardo Dalla-Favera, Andrea Califano&lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': gene expression, transcriptional interaction, modulator &lt;br /&gt;
* '''URL''': n/a&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;br /&gt;
&lt;br /&gt;
===B Cell Interactome===&lt;br /&gt;
* '''Description''': The B cell interactome (BCI)  is a network of protein-protein, protein-DNA and modulatory interactions in human B cells. The network contains known interactions (reported in public databases) and predicted interactions by a Bayesian evidence integration framework which integrates a variety of generic and context specific experimental clues about protein-protein and protein-DNA interactions - such as a large collection of B cell expression profiles - with inferences from different reverse engineering algorithms, such as GeneWays and ARACNE. Modulatory interactions are predicted by MINDY, an algorithm for the prediction of modulators of transcriptional interactions.&lt;br /&gt;
* '''Data Input''': n/a&lt;br /&gt;
* '''Data Output''':  text file of binary interations associated with a probability. &lt;br /&gt;
* '''Implementation Language''': Perl&lt;br /&gt;
* '''Version, Date, Stage''': Version 2, March 2007&lt;br /&gt;
* '''Authors''':  Lefebvre C, Lim WK, Basso K, Dalla Favera R, and Califano A.&lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': Naive Bayes, Mixed-Interaction Network, human B cells.&lt;br /&gt;
* '''URL''': http://amdec-bioinfo.cu-genome.org/html/BCellInteractome.html&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Interaction Modeling&lt;br /&gt;
&lt;br /&gt;
===ARACNE===&lt;br /&gt;
* '''Description''': ARACNE is an algorithm for inferring gene regulatory networks from a set of microarray experiments. The method uses mutual information to identify genes that are co-expressed and then applies the data processing inequality to filter out interactions that are likely to be indirect.&lt;br /&gt;
* '''Data Input''': Text file containing measurements from a set of microarray experiments.&lt;br /&gt;
* '''Data Output''':  Text file containing predicted interactions.&lt;br /&gt;
* '''Implementation Language''': C++, Java&lt;br /&gt;
* '''Version, Date, Stage''': Version 1, June, 2006&lt;br /&gt;
* '''Authors''':  Margolin AA, Nemenman I, Basso K, Wiggins C, Stolovitzky G, Dalla Favera R, Califano A.&lt;br /&gt;
* '''Platforms Tested''': Window, Linux&lt;br /&gt;
* '''License''':  Open source&lt;br /&gt;
* '''Keywords''': Reverse engineering, mutual information, genetic networks, microarray&lt;br /&gt;
* '''URL''': http://amdec-bioinfo.cu-genome.org/html/ARACNE.htm&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===geWorkbench===&lt;br /&gt;
* '''Description''': geWorkbench is a Java application that provides users with an integrated suite of genomics tools. It is built on an open-source, extensible architecture that promotes interoperability and simplifies the development of new as well as the incorporation of pre-existing components. The resulting system provides seamless access to a multitude of both local and remote data and computational services through an integrated environment that offers a unified user experience. Over 50 data analysis and visualization components have been developed for the framework, covering a wide range of genomics domains including gene expression, sequence, structure and network data.&lt;br /&gt;
* '''Data Input''': Gene epxression data (Affy, GenPix, RMA), Sequence (FASTA), Structure (PDB).&lt;br /&gt;
* '''Data Output''': Analysis results (multiple formats).&lt;br /&gt;
* '''Implementation Language''': Java&lt;br /&gt;
* '''Version, Date, Stage''': 1.0.5, 3/23/07, stable production release&lt;br /&gt;
* '''Authors''':  A. Califano, A. Floratos. M. Kustagi, K. Smith, J. Watkinson, M. Hall, K. Keshav, X. Zhang, K. Kushal, B. Jagla, E. Daly, M. VanGinhoven, P. Morozov.&lt;br /&gt;
* '''Platforms Tested''': Windows XP, Linux, Mac OS 10.x.&lt;br /&gt;
* '''License''': Free. &lt;br /&gt;
* '''Keywords''': Analysis suite, gene expression analysis, sequence analysis, network reconstruction, structure predcition, visualization.&lt;br /&gt;
* '''URL''': http://www.geworkbench.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Interaction Modeling; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Software Engineering and Development Tool --&amp;gt; Integration --&amp;gt; Resource Integration Components; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Software Engineering and Development Tool --&amp;gt; Integration --&amp;gt; Grid Computing Resources; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Visualization&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10272</id>
		<title>SDIWG:NCBC Software Classification MAGNet Examples</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10272"/>
		<updated>2007-05-13T22:25:54Z</updated>

		<summary type="html">&lt;p&gt;Aris: /* GRASP2 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;===DelPhi===&lt;br /&gt;
* '''Description''': DelPhi provides numerical solutions to the Poisson-Boltzmann equation (both linear and nonlinear form) for molecules of arbitrary shape and charge distribution. The current version is fast (the best relaxation parameter is estimated at run time), accurate (calculation of the electrostatic free energy is less dependent on the resolution of the lattice) and can handle extremely high lattice dimensions. It also includes flexible features for assigning different dielectric constants to different regions of space and treating systems containing mixed salt solutions.&lt;br /&gt;
* '''Data Input''': DelPhi takes as input a coordinate file format of a molecule or equivalent data for geometrical objects and/or charge distributions&lt;br /&gt;
* '''Data Output''': electrostatic potential in and around the system&lt;br /&gt;
* '''Implementation Language''': Fortran and C&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''':  E.Alexov,  R.Fine, M.K.Gilson, A.Nicholls, W.Rocchia, K.Sharp, and B. Honig.&lt;br /&gt;
* '''Platforms Tested''': Unix-SGI IRIX, linux, PC (requires Fortran and C compilers), AIX IBM version and Mac.&lt;br /&gt;
* '''License''': Freely available to academia; pay model for commercial users. &lt;br /&gt;
* '''Keywords''': Finite Difference Poisson-Boltzman Solver&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/delphi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Numerical Calculation of Electrostatic Potential&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP===&lt;br /&gt;
* '''Description''': A molecular visualization and analysis program. It is particularly useful for the display and manipulation of the surfaces of molecules and their electrostatic properties.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi&lt;br /&gt;
* '''Data Output''': molecular graphics.&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': v1.3.6 .Stable public release.&lt;br /&gt;
* '''Authors''': Anthony Nicholls and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': SGI machines: irix 5.x and 6.x (INDYs, INDIGOs including Impact, Octane and O2) systems.&lt;br /&gt;
* '''License''':  Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
===Nest===&lt;br /&gt;
* '''Description''': Modeling protein structure based on a sequence-template alignment. The current server works only for modeling with a single template. Part of jackal, which can be downloaded.&lt;br /&gt;
* '''Data Input''': pir and PDB files&lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Xiang, Z. and Honig, B.&lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': modeling, protein structure, sequence-template alignment.&lt;br /&gt;
* '''URL''': http://honiglab.cpmc.columbia.edu/cgi-bin/jackal/nest.cgi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===JACKAL===&lt;br /&gt;
* '''Description''': Jackal is a collection of programs designed for the modeling and analysis of protein structures. Its core program is a versatile homology modeling package nest. JACKAL has the following capabilities: 1) comparative modeling based on single, composite or multiple templates; 2) side-chain prediction; 3) modeling residue mutation, insertion or deletion; 4) loop prediction; 5) structure refinement; 6) reconstruction of protein missing atoms;7) reconstruction of protein missing residues; 8) prediction of hydrogen atoms; 9) fast calculation of solvent accessible surface area; 10) structure superimposition.       &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Version: 1.5 as of Oct, 20, 2002, Stable public release.&lt;br /&gt;
* '''Authors''': Z. Xiang and B. Honig&lt;br /&gt;
* '''Platforms Tested''': SGI 6.5, Intel Linux and Sun solaris&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': Protein Structure Modeling&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/jackal&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP2===&lt;br /&gt;
* '''Description''': GRASP2 is an updated version of the GRASP program used for macromolecular structure and surface visualization, contains a large number of new features and scientific tools: Enhanced GUI; Structure alignment and domain database scanning; A gaussian surface generator and new surface coloring schemes; Sequence visualization and alignment; Completed work can be stored in &amp;quot;project files;  Among the many objects that can be stored in a project file are views of the structure; defined subsets, surfaces; Direct printing to printers at full printer resolution.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi, sequence alignments.&lt;br /&gt;
* '''Data Output''': molecular graphics, structural alignments.&lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Donald Petrey and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp2&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
===PrISM===&lt;br /&gt;
* '''Description''': PrISM is an integrated computational system where computational tools are implemented for protein sequence and structure analysis and modeling.&lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Wang, L, Yang, A. S. &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''':SGI-irix, Intel-linux &lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': protein analysis/modeling &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/PrISM/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Protein-DNA interface alignment===&lt;br /&gt;
* '''Description''': The protein-DNA alignment software allows one to align the interfacial amino acids from two protein-DNA complexes based on the geometric relationship of each amino acid to its local DNA. &lt;br /&gt;
* '''Data Input''': two PDB files that both contain protein-DNA complexes&lt;br /&gt;
* '''Data Output''': The programs will output the aligned residues and their corresponding residue-residue similarity scores, s(i,j).&lt;br /&gt;
* '''Implementation Language''': C++ and Perl&lt;br /&gt;
* '''Version, Date, Stage''':Stable public release. &lt;br /&gt;
* '''Authors''': Siggers, T.W., Silkov, A &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''': Linux&lt;br /&gt;
* '''License''': Freely available to academia &lt;br /&gt;
* '''Keywords''': protein-DNA interface &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/intfc_aln&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
===SURFace===&lt;br /&gt;
* '''Description''': SURFace algorithms are programs that calculate solvent accessible surface area and curvature corrected solvent accessible surface area &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': &lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Nicholls, A., Sharp, K., Sridharan, S. and Honig, B.  &lt;br /&gt;
* '''Platforms Tested''': SGI&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': solvent accessible surface area  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/surf/&lt;br /&gt;
* '''Organization''': MAGNet &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Caculation of Solvent Accessible Area&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Target Explorer===&lt;br /&gt;
* '''Description''': Automated process of prediction of complex regulatory elements for specified set of transcription factors in Drosophila melanogaster genome. Target Explorer is a complex tool with user-friendly self-explanatory Web-interface that allows to user:  1. create customized library of TF binding site matrices based on user defined sets of training sequences; 2. search for new clusters of binding sites for specified set of TFs; 3.extract annotation for potential target genes.    &lt;br /&gt;
* '''Data Input''': genomic sequences&lt;br /&gt;
* '''Data Output''': clusters of known binding sites &lt;br /&gt;
* '''Implementation Language''': perl, cgi&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release. &lt;br /&gt;
* '''Authors''': Sosinsky A, Bonin CP, Mann RS, Honig B.   &lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': prediction of binding sites for transcription factors  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/Target_Explorer/&lt;br /&gt;
* '''Organization''': MAGNet  &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis--&amp;gt; Sequence Annotation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===MEDUSA and Gorgon===&lt;br /&gt;
* '''Description''': MEDUSA is an algorithm for learning predictive models of transcriptional gene regulation from gene expression and promoter sequence data.  By using a statistical learning approach based on boosing, MEDUSA learns cis regulatory motifs, condition-specific regulators, and regulatory programs that predict the differential expression of target genes.  The regulatory program is specified as an alternating decision tree (ADT).  The Java implementation of MEDUSA will allow a number of visualizations of the regulatory program and other inferred regulatory information, implemented in the accompanying Gorgon tool, including hits of significant and condition-specific motifs along the promoter sequences of target genes and regulatory network figures viewable in Cytoscape.      &lt;br /&gt;
* '''Data Input''': Discretized (up/down/baseline) gene expression data in plain text format, promoter sequences in FASTA format, list of candidate transcriptional regulators and signal transducers in plain text format.&lt;br /&gt;
* '''Data Output''': Regulatory program represented as a Java serialized object file readable by Gorgon and as a human readable XML file.  Gorgon currently generates views of learned PSSMs, positional hits along promoter sequences, and views of the ADT as HTML files, and generates network figures as Cytoscape format files.&lt;br /&gt;
* '''Implementation Language''': Java (prototyped in MATLAB) &lt;br /&gt;
* '''Version, Date, Stage''': Version 2.0, July 2006, pre-release beta version; Version 1.0 (MATLAB), April 2005, stable public release  &lt;br /&gt;
* '''Authors''': David Quigley, Manuel Middendorf, Steve Lianoglou, Anshul Kundaje, Yoav Freund, Chris Wiggins, Christina Leslie   &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux, Mac OS X&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/medusa (MATLAB),http://compbio.sytes.net:8090/medusa (Java beta version)&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===String kernel package===&lt;br /&gt;
* '''Description''':The string kernel package contains implementations for the mismatch and profile string kernels for use with support vector machine (SVM) classifiers for protein sequence classification.  Both kernels compute similarity between protein sequences based on common occurrences of k-length subsequences (&amp;quot;k-mers&amp;quot;) counted with substitutions.  Kernel functions for protein sequence data enable the training of SVMs for a range of prediction problems, in particular protein structural class prediction and remote homology detection.  A version of the Spider MATLAB machine learning package is also bundled with the code, which allows users to train SVMs and evaluate performance on test sets with the packaged software.&lt;br /&gt;
* '''Data Input''': The mismatch kernel requires sequence data in FASTA format. The profile string kernel uses probabilistic profiles, such as those produced by PSI-BLAST, in place of the original sequences.  The Spider SVM implementation requires both the kernel matrix and a label file of binary or multi-class labels for the training data; this data must be loaded into MATLAB variables before using Spider routing.&lt;br /&gt;
* '''Data Output''':The kernel code produces a kernel matrix for the input data in tab-delimited text format.  The Spider package trains SVMs and stores the learns classifier and results from applying the classifier on test data as MATLAB objects.&lt;br /&gt;
* '''Implementation Language''': String kernel code is implemented in C. Spider is a set of object-oriented MATLAB routines. &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.2, September 2004, stable public release &lt;br /&gt;
* '''Authors''': Eleazar Eskin, Rui Kuang, Eugene Ie, Ke Wang, Jason Weston, Bill Noble, Christina Leslie  &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/string-kernels&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MatrixREDUCE===&lt;br /&gt;
* '''Description''': Regulation of gene expression by a transcription factor requires physical interaction between the factor and the DNA, which can be described by astatistical mechanical model. Based on this model, the MatrixREDUCE algorithm uses genome-wide occupancy data for a transcription factor (e.g.ChIP-chip or mRNA expression data) and associated nucleotide sequences to discover the sequence-specific binding affinity of the transcription factor. The sequence specificity of the transcription factor's DNA-binding domain is modeled using a position-specific affinity matrix (PSAM), representing the change in the binding affinity (Kd) whenever a specific position within a reference binding sequence is mutated. The PSAM can be transformed into affinity logo for visualization using the utility program AffinityLogo, and a MatrixREDUCE run can be summarized in an easy-to-navigate webpage using HTMLSummary. &lt;br /&gt;
* '''Data Input''': sequence file in FASTA format; and expression data file in tab-delimited text format.&lt;br /&gt;
* '''Data Output''': PSAMs in numeric and graphical format, parameters of the fitted model, and an HTML summary page.&lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of Numerical Recipes routines.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.0, July 10, 2006, extensively tested in lab.  &lt;br /&gt;
* '''Authors''': Barrett Foat, Xiang-Jun Lu, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': position-specific affinity matrix, binding affinity, cis-regulatory element, expression data, ChIP-chip, transcription factor &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/MatrixREDUCE&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===T-profiler===&lt;br /&gt;
* '''Description''': T-profiler is a web-based tool that uses the t-test to score changes in the average activity of pre-defined groups of genes. The gene groups are defined based on Gene Ontology categorization, ChIP-chip experiments, upstream matches to a consensus transcription factor binding motif, and location on the same chromosome, respectively. If desired, an iterative procedure can be used to select a single, optimal representative from sets of overlapping gene groups. A jack-knife procedure is used to make calculations more robust against outliers. T-profiler makes it possible to interpret microarray data in a way that is both intuitive and statistically rigorous, without the need to combine experiments or choose parameters.   &lt;br /&gt;
* '''Data Input''': Currently, gene expression data from Saccharomyces cerevisiae and Candida albicans are supported. &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': T-profiler is written in PHP, data is managed by a MYSQL database server  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': André Boorsma, Barrett C. Foat, Daniel Vis, Frans Klis, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Web-based application&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': gene expression, transcriptome,  ChIP-chip,  Gene Ontology &lt;br /&gt;
* '''URL''': http://www.t-profiler.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===TranscriptionDetector===&lt;br /&gt;
* '''Description''':A tool for finding probes measuring significantly expressed loci in a genomic array experiment. Given expression data from some tiling array experiment, TranscriptionDetector decides the likelihood that a probe is detecting transcription from the locus in which it resides. Probabilities are assigned by making use of a background signal intensity distribution from a set of negative control probes. This tool is useful for the functional annotation of genomes as it allows for the discovery of novel transcriptional units independently of any genomic annotation.    &lt;br /&gt;
* '''Data Input''': Expression data (GEO or other platforms) and designation of which probes represent negative controls and which are data probes. &lt;br /&gt;
* '''Data Output''': A text file with a list of probes  corresponding to significantly expressed loci. &lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of GSL.  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': Xiang-Jun Lu, Gabor Halasz, Marinus F. van Batenburg  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': tiling arrays, expression, transcriptome &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/TranscriptionDetector/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===PhenoGO===&lt;br /&gt;
* '''Description''': PhenoGO adds phenotypic contextual information to existing associations between gene products and Gene Ontology (GO) terms as specified in GO Annotations (GOA). PhenoGO utilizes an existing Natural Language Processing (NLP) system, called BioMedLEE, an existing knowledge-based phenotype organizer system (PhenOS) in conjunction with MeSH indexing and established biomedical ontologies. The system also encodes the context to identifiers that are associated in different biomedical ontologies, including the UMLS, Cell Ontology, Mouse Anatomy, NCBI taxonomy, GO, and Mammalian Phenotype Ontology. In addition, PhenoGO was evaluated for coding of anatomical and cellular information and assigning the coded phenotypes to the correct GOA; results obtained show that PhenoGO has a precision of 91% and recall of 92%, demonstrating that the PhenoGO NLP system can accurately encode a large number of anatomical and cellular ontologies to GO annotations. The PhenoGO Database may be accessed at www.phenogo.org.    &lt;br /&gt;
* '''Data Input''': Gene Ontology Annotations Files and Medline Abstracts  &lt;br /&gt;
* '''Data Output''': XML file and www.phenogo.org Web Portal&lt;br /&gt;
* '''Implementation Language''': A variety of modules, the web portal is in Java and MySQL, the computational terminology component (phenOS) is written in Perl scripts that queries tables in IBM DB2,  the natural language processing component is written in PROLOG.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 2, Feb 2006  &lt;br /&gt;
* '''Authors''': Yves Lussier and Carol Friedman are the principal investigators. The programmers are Jianrong Li, Lee Sam, and Tara Borlawsky   &lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': Phenotypic integration, computational phenotypes &lt;br /&gt;
* '''URL''': www.phenogo.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MINDY===&lt;br /&gt;
* '''Description''': Given a transcription factor of interest, MINDY uses a large set of gene expression profile data to identify potential post-transcriptional modulators of the transcription factor's activity. MINDY is based on a three-way statistical interaction model that captures the post-transcriptional regulatory event where the ability of a transcription factor to activate/repress its target genes is monotonically controlled by a potential modulator gene. &lt;br /&gt;
* '''Data Input''': Gene expression data in the EXP format, and a user-specified transcription factor of interest&lt;br /&gt;
* '''Data Output''': Lists of the putative modulators and target genes of the transcription factor, and the modulatory interactions involving them&lt;br /&gt;
* '''Implementation Language''': C++ and MATLAB, Java&lt;br /&gt;
* '''Version, Date, Stage''': Stable release, April 2007  &lt;br /&gt;
* '''Authors''': Kai Wang, Ilya Nemenman, Adam Margolin, Riccardo Dalla-Favera, Andrea Califano&lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': gene expression, transcriptional interaction, modulator &lt;br /&gt;
* '''URL''': n/a&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;br /&gt;
&lt;br /&gt;
===B Cell Interactome===&lt;br /&gt;
* '''Description''': The B cell interactome (BCI)  is a network of protein-protein, protein-DNA and modulatory interactions in human B cells. The network contains known interactions (reported in public databases) and predicted interactions by a Bayesian evidence integration framework which integrates a variety of generic and context specific experimental clues about protein-protein and protein-DNA interactions - such as a large collection of B cell expression profiles - with inferences from different reverse engineering algorithms, such as GeneWays and ARACNE. Modulatory interactions are predicted by MINDY, an algorithm for the prediction of modulators of transcriptional interactions.&lt;br /&gt;
* '''Data Input''': n/a&lt;br /&gt;
* '''Data Output''':  text file of binary interations associated with a probability. &lt;br /&gt;
* '''Implementation Language''': Perl&lt;br /&gt;
* '''Version, Date, Stage''': Version 2, March 2007&lt;br /&gt;
* '''Authors''':  Lefebvre C, Lim WK, Basso K, Dalla Favera R, and Califano A.&lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': Naive Bayes, Mixed-Interaction Network, human B cells.&lt;br /&gt;
* '''URL''': http://amdec-bioinfo.cu-genome.org/html/BCellInteractome.html&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Interaction Modeling&lt;br /&gt;
&lt;br /&gt;
===ARACNE===&lt;br /&gt;
* '''Description''': ARACNE is an algorithm for inferring gene regulatory networks from a set of microarray experiments. The method uses mutual information to identify genes that are co-expressed and then applies the data processing inequality to filter out interactions that are likely to be indirect.&lt;br /&gt;
* '''Data Input''': Text file containing measurements from a set of microarray experiments.&lt;br /&gt;
* '''Data Output''':  Text file containing predicted interactions.&lt;br /&gt;
* '''Implementation Language''': C++, Java&lt;br /&gt;
* '''Version, Date, Stage''': Version 1, June, 2006&lt;br /&gt;
* '''Authors''':  Margolin AA, Nemenman I, Basso K, Wiggins C, Stolovitzky G, Dalla Favera R, Califano A.&lt;br /&gt;
* '''Platforms Tested''': Window, Linux&lt;br /&gt;
* '''License''':  Open source&lt;br /&gt;
* '''Keywords''': Reverse engineering, mutual information, genetic networks, microarray&lt;br /&gt;
* '''URL''': http://amdec-bioinfo.cu-genome.org/html/ARACNE.htm&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===geWorkbench===&lt;br /&gt;
* '''Description''': geWorkbench is a Java application that provides users with an integrated suite of genomics tools. It is built on an open-source, extensible architecture that promotes interoperability and simplifies the development of new as well as the incorporation of pre-existing components. The resulting system provides seamless access to a multitude of both local and remote data and computational services through an integrated environment that offers a unified user experience. Over 50 data analysis and visualization components have been developed for the framework, covering a wide range of genomics domains including gene expression, sequence, structure and network data.&lt;br /&gt;
* '''Data Input''': Gene epxression data (Affy, GenPix, RMA), Sequence (FASTA), Structure (PDB).&lt;br /&gt;
* '''Data Output''': Analysis results (multiple formats).&lt;br /&gt;
* '''Implementation Language''': Java&lt;br /&gt;
* '''Version, Date, Stage''': 1.0.5, 3/23/07, stable production release&lt;br /&gt;
* '''Authors''':  A. Califano, A. Floratos. M. Kustagi, K. Smith, J. Watkinson, M. Hall, K. Keshav, X. Zhang, K. Kushal, B. Jagla, E. Daly, M. VanGinhoven, P. Morozov.&lt;br /&gt;
* '''Platforms Tested''': Windows XP, Linux, Mac OS 10.x.&lt;br /&gt;
* '''License''': Free. &lt;br /&gt;
* '''Keywords''': Analysis suite, gene expression analysis, sequence analysis, network reconstruction, structure predcition, visualization.&lt;br /&gt;
* '''URL''': http://www.geworkbench.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Interaction Modeling; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Software Engineering and Development Tool --&amp;gt; Integration --&amp;gt; Resource Integration Components; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Software Engineering and Development Tool --&amp;gt; Integration --&amp;gt; Grid Computing Resources; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Visualization&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10271</id>
		<title>SDIWG:NCBC Software Classification MAGNet Examples</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10271"/>
		<updated>2007-05-13T22:23:35Z</updated>

		<summary type="html">&lt;p&gt;Aris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;===DelPhi===&lt;br /&gt;
* '''Description''': DelPhi provides numerical solutions to the Poisson-Boltzmann equation (both linear and nonlinear form) for molecules of arbitrary shape and charge distribution. The current version is fast (the best relaxation parameter is estimated at run time), accurate (calculation of the electrostatic free energy is less dependent on the resolution of the lattice) and can handle extremely high lattice dimensions. It also includes flexible features for assigning different dielectric constants to different regions of space and treating systems containing mixed salt solutions.&lt;br /&gt;
* '''Data Input''': DelPhi takes as input a coordinate file format of a molecule or equivalent data for geometrical objects and/or charge distributions&lt;br /&gt;
* '''Data Output''': electrostatic potential in and around the system&lt;br /&gt;
* '''Implementation Language''': Fortran and C&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''':  E.Alexov,  R.Fine, M.K.Gilson, A.Nicholls, W.Rocchia, K.Sharp, and B. Honig.&lt;br /&gt;
* '''Platforms Tested''': Unix-SGI IRIX, linux, PC (requires Fortran and C compilers), AIX IBM version and Mac.&lt;br /&gt;
* '''License''': Freely available to academia; pay model for commercial users. &lt;br /&gt;
* '''Keywords''': Finite Difference Poisson-Boltzman Solver&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/delphi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Numerical Calculation of Electrostatic Potential&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP===&lt;br /&gt;
* '''Description''': A molecular visualization and analysis program. It is particularly useful for the display and manipulation of the surfaces of molecules and their electrostatic properties.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi&lt;br /&gt;
* '''Data Output''': molecular graphics.&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': v1.3.6 .Stable public release.&lt;br /&gt;
* '''Authors''': Anthony Nicholls and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': SGI machines: irix 5.x and 6.x (INDYs, INDIGOs including Impact, Octane and O2) systems.&lt;br /&gt;
* '''License''':  Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
===Nest===&lt;br /&gt;
* '''Description''': Modeling protein structure based on a sequence-template alignment. The current server works only for modeling with a single template. Part of jackal, which can be downloaded.&lt;br /&gt;
* '''Data Input''': pir and PDB files&lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Xiang, Z. and Honig, B.&lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': modeling, protein structure, sequence-template alignment.&lt;br /&gt;
* '''URL''': http://honiglab.cpmc.columbia.edu/cgi-bin/jackal/nest.cgi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===JACKAL===&lt;br /&gt;
* '''Description''': Jackal is a collection of programs designed for the modeling and analysis of protein structures. Its core program is a versatile homology modeling package nest. JACKAL has the following capabilities: 1) comparative modeling based on single, composite or multiple templates; 2) side-chain prediction; 3) modeling residue mutation, insertion or deletion; 4) loop prediction; 5) structure refinement; 6) reconstruction of protein missing atoms;7) reconstruction of protein missing residues; 8) prediction of hydrogen atoms; 9) fast calculation of solvent accessible surface area; 10) structure superimposition.       &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Version: 1.5 as of Oct, 20, 2002, Stable public release.&lt;br /&gt;
* '''Authors''': Z. Xiang and B. Honig&lt;br /&gt;
* '''Platforms Tested''': SGI 6.5, Intel Linux and Sun solaris&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': Protein Structure Modeling&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/jackal&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP2===&lt;br /&gt;
* '''Description''': GRASP2 is an updated version of the GRASP program used for macromolecular structure and surface visualization, contains a large number of new features and scientific tools: Enhanced GUI; Structure alignment and domain database scanning; A gaussian surface generator and new surface coloring schemes; Sequence visualization and alignment; Completed work can be stored in &amp;quot;project files;  Among the many objects that can be stored in a project file are views of the structure; defined subsets, surfaces; Direct printing to printers at full printer resolution.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi, sequence alignments.&lt;br /&gt;
* '''Data Output''': molecular graphics, structural alignments.&lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Donald Petrey and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp2&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===PrISM===&lt;br /&gt;
* '''Description''': PrISM is an integrated computational system where computational tools are implemented for protein sequence and structure analysis and modeling.&lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Wang, L, Yang, A. S. &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''':SGI-irix, Intel-linux &lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': protein analysis/modeling &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/PrISM/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Protein-DNA interface alignment===&lt;br /&gt;
* '''Description''': The protein-DNA alignment software allows one to align the interfacial amino acids from two protein-DNA complexes based on the geometric relationship of each amino acid to its local DNA. &lt;br /&gt;
* '''Data Input''': two PDB files that both contain protein-DNA complexes&lt;br /&gt;
* '''Data Output''': The programs will output the aligned residues and their corresponding residue-residue similarity scores, s(i,j).&lt;br /&gt;
* '''Implementation Language''': C++ and Perl&lt;br /&gt;
* '''Version, Date, Stage''':Stable public release. &lt;br /&gt;
* '''Authors''': Siggers, T.W., Silkov, A &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''': Linux&lt;br /&gt;
* '''License''': Freely available to academia &lt;br /&gt;
* '''Keywords''': protein-DNA interface &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/intfc_aln&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
===SURFace===&lt;br /&gt;
* '''Description''': SURFace algorithms are programs that calculate solvent accessible surface area and curvature corrected solvent accessible surface area &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': &lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Nicholls, A., Sharp, K., Sridharan, S. and Honig, B.  &lt;br /&gt;
* '''Platforms Tested''': SGI&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': solvent accessible surface area  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/surf/&lt;br /&gt;
* '''Organization''': MAGNet &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Caculation of Solvent Accessible Area&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Target Explorer===&lt;br /&gt;
* '''Description''': Automated process of prediction of complex regulatory elements for specified set of transcription factors in Drosophila melanogaster genome. Target Explorer is a complex tool with user-friendly self-explanatory Web-interface that allows to user:  1. create customized library of TF binding site matrices based on user defined sets of training sequences; 2. search for new clusters of binding sites for specified set of TFs; 3.extract annotation for potential target genes.    &lt;br /&gt;
* '''Data Input''': genomic sequences&lt;br /&gt;
* '''Data Output''': clusters of known binding sites &lt;br /&gt;
* '''Implementation Language''': perl, cgi&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release. &lt;br /&gt;
* '''Authors''': Sosinsky A, Bonin CP, Mann RS, Honig B.   &lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': prediction of binding sites for transcription factors  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/Target_Explorer/&lt;br /&gt;
* '''Organization''': MAGNet  &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis--&amp;gt; Sequence Annotation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===MEDUSA and Gorgon===&lt;br /&gt;
* '''Description''': MEDUSA is an algorithm for learning predictive models of transcriptional gene regulation from gene expression and promoter sequence data.  By using a statistical learning approach based on boosing, MEDUSA learns cis regulatory motifs, condition-specific regulators, and regulatory programs that predict the differential expression of target genes.  The regulatory program is specified as an alternating decision tree (ADT).  The Java implementation of MEDUSA will allow a number of visualizations of the regulatory program and other inferred regulatory information, implemented in the accompanying Gorgon tool, including hits of significant and condition-specific motifs along the promoter sequences of target genes and regulatory network figures viewable in Cytoscape.      &lt;br /&gt;
* '''Data Input''': Discretized (up/down/baseline) gene expression data in plain text format, promoter sequences in FASTA format, list of candidate transcriptional regulators and signal transducers in plain text format.&lt;br /&gt;
* '''Data Output''': Regulatory program represented as a Java serialized object file readable by Gorgon and as a human readable XML file.  Gorgon currently generates views of learned PSSMs, positional hits along promoter sequences, and views of the ADT as HTML files, and generates network figures as Cytoscape format files.&lt;br /&gt;
* '''Implementation Language''': Java (prototyped in MATLAB) &lt;br /&gt;
* '''Version, Date, Stage''': Version 2.0, July 2006, pre-release beta version; Version 1.0 (MATLAB), April 2005, stable public release  &lt;br /&gt;
* '''Authors''': David Quigley, Manuel Middendorf, Steve Lianoglou, Anshul Kundaje, Yoav Freund, Chris Wiggins, Christina Leslie   &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux, Mac OS X&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/medusa (MATLAB),http://compbio.sytes.net:8090/medusa (Java beta version)&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===String kernel package===&lt;br /&gt;
* '''Description''':The string kernel package contains implementations for the mismatch and profile string kernels for use with support vector machine (SVM) classifiers for protein sequence classification.  Both kernels compute similarity between protein sequences based on common occurrences of k-length subsequences (&amp;quot;k-mers&amp;quot;) counted with substitutions.  Kernel functions for protein sequence data enable the training of SVMs for a range of prediction problems, in particular protein structural class prediction and remote homology detection.  A version of the Spider MATLAB machine learning package is also bundled with the code, which allows users to train SVMs and evaluate performance on test sets with the packaged software.&lt;br /&gt;
* '''Data Input''': The mismatch kernel requires sequence data in FASTA format. The profile string kernel uses probabilistic profiles, such as those produced by PSI-BLAST, in place of the original sequences.  The Spider SVM implementation requires both the kernel matrix and a label file of binary or multi-class labels for the training data; this data must be loaded into MATLAB variables before using Spider routing.&lt;br /&gt;
* '''Data Output''':The kernel code produces a kernel matrix for the input data in tab-delimited text format.  The Spider package trains SVMs and stores the learns classifier and results from applying the classifier on test data as MATLAB objects.&lt;br /&gt;
* '''Implementation Language''': String kernel code is implemented in C. Spider is a set of object-oriented MATLAB routines. &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.2, September 2004, stable public release &lt;br /&gt;
* '''Authors''': Eleazar Eskin, Rui Kuang, Eugene Ie, Ke Wang, Jason Weston, Bill Noble, Christina Leslie  &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/string-kernels&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MatrixREDUCE===&lt;br /&gt;
* '''Description''': Regulation of gene expression by a transcription factor requires physical interaction between the factor and the DNA, which can be described by astatistical mechanical model. Based on this model, the MatrixREDUCE algorithm uses genome-wide occupancy data for a transcription factor (e.g.ChIP-chip or mRNA expression data) and associated nucleotide sequences to discover the sequence-specific binding affinity of the transcription factor. The sequence specificity of the transcription factor's DNA-binding domain is modeled using a position-specific affinity matrix (PSAM), representing the change in the binding affinity (Kd) whenever a specific position within a reference binding sequence is mutated. The PSAM can be transformed into affinity logo for visualization using the utility program AffinityLogo, and a MatrixREDUCE run can be summarized in an easy-to-navigate webpage using HTMLSummary. &lt;br /&gt;
* '''Data Input''': sequence file in FASTA format; and expression data file in tab-delimited text format.&lt;br /&gt;
* '''Data Output''': PSAMs in numeric and graphical format, parameters of the fitted model, and an HTML summary page.&lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of Numerical Recipes routines.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.0, July 10, 2006, extensively tested in lab.  &lt;br /&gt;
* '''Authors''': Barrett Foat, Xiang-Jun Lu, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': position-specific affinity matrix, binding affinity, cis-regulatory element, expression data, ChIP-chip, transcription factor &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/MatrixREDUCE&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===T-profiler===&lt;br /&gt;
* '''Description''': T-profiler is a web-based tool that uses the t-test to score changes in the average activity of pre-defined groups of genes. The gene groups are defined based on Gene Ontology categorization, ChIP-chip experiments, upstream matches to a consensus transcription factor binding motif, and location on the same chromosome, respectively. If desired, an iterative procedure can be used to select a single, optimal representative from sets of overlapping gene groups. A jack-knife procedure is used to make calculations more robust against outliers. T-profiler makes it possible to interpret microarray data in a way that is both intuitive and statistically rigorous, without the need to combine experiments or choose parameters.   &lt;br /&gt;
* '''Data Input''': Currently, gene expression data from Saccharomyces cerevisiae and Candida albicans are supported. &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': T-profiler is written in PHP, data is managed by a MYSQL database server  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': André Boorsma, Barrett C. Foat, Daniel Vis, Frans Klis, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Web-based application&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': gene expression, transcriptome,  ChIP-chip,  Gene Ontology &lt;br /&gt;
* '''URL''': http://www.t-profiler.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===TranscriptionDetector===&lt;br /&gt;
* '''Description''':A tool for finding probes measuring significantly expressed loci in a genomic array experiment. Given expression data from some tiling array experiment, TranscriptionDetector decides the likelihood that a probe is detecting transcription from the locus in which it resides. Probabilities are assigned by making use of a background signal intensity distribution from a set of negative control probes. This tool is useful for the functional annotation of genomes as it allows for the discovery of novel transcriptional units independently of any genomic annotation.    &lt;br /&gt;
* '''Data Input''': Expression data (GEO or other platforms) and designation of which probes represent negative controls and which are data probes. &lt;br /&gt;
* '''Data Output''': A text file with a list of probes  corresponding to significantly expressed loci. &lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of GSL.  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': Xiang-Jun Lu, Gabor Halasz, Marinus F. van Batenburg  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': tiling arrays, expression, transcriptome &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/TranscriptionDetector/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===PhenoGO===&lt;br /&gt;
* '''Description''': PhenoGO adds phenotypic contextual information to existing associations between gene products and Gene Ontology (GO) terms as specified in GO Annotations (GOA). PhenoGO utilizes an existing Natural Language Processing (NLP) system, called BioMedLEE, an existing knowledge-based phenotype organizer system (PhenOS) in conjunction with MeSH indexing and established biomedical ontologies. The system also encodes the context to identifiers that are associated in different biomedical ontologies, including the UMLS, Cell Ontology, Mouse Anatomy, NCBI taxonomy, GO, and Mammalian Phenotype Ontology. In addition, PhenoGO was evaluated for coding of anatomical and cellular information and assigning the coded phenotypes to the correct GOA; results obtained show that PhenoGO has a precision of 91% and recall of 92%, demonstrating that the PhenoGO NLP system can accurately encode a large number of anatomical and cellular ontologies to GO annotations. The PhenoGO Database may be accessed at www.phenogo.org.    &lt;br /&gt;
* '''Data Input''': Gene Ontology Annotations Files and Medline Abstracts  &lt;br /&gt;
* '''Data Output''': XML file and www.phenogo.org Web Portal&lt;br /&gt;
* '''Implementation Language''': A variety of modules, the web portal is in Java and MySQL, the computational terminology component (phenOS) is written in Perl scripts that queries tables in IBM DB2,  the natural language processing component is written in PROLOG.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 2, Feb 2006  &lt;br /&gt;
* '''Authors''': Yves Lussier and Carol Friedman are the principal investigators. The programmers are Jianrong Li, Lee Sam, and Tara Borlawsky   &lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': Phenotypic integration, computational phenotypes &lt;br /&gt;
* '''URL''': www.phenogo.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MINDY===&lt;br /&gt;
* '''Description''': Given a transcription factor of interest, MINDY uses a large set of gene expression profile data to identify potential post-transcriptional modulators of the transcription factor's activity. MINDY is based on a three-way statistical interaction model that captures the post-transcriptional regulatory event where the ability of a transcription factor to activate/repress its target genes is monotonically controlled by a potential modulator gene. &lt;br /&gt;
* '''Data Input''': Gene expression data in the EXP format, and a user-specified transcription factor of interest&lt;br /&gt;
* '''Data Output''': Lists of the putative modulators and target genes of the transcription factor, and the modulatory interactions involving them&lt;br /&gt;
* '''Implementation Language''': C++ and MATLAB, Java&lt;br /&gt;
* '''Version, Date, Stage''': Stable release, April 2007  &lt;br /&gt;
* '''Authors''': Kai Wang, Ilya Nemenman, Adam Margolin, Riccardo Dalla-Favera, Andrea Califano&lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': gene expression, transcriptional interaction, modulator &lt;br /&gt;
* '''URL''': n/a&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;br /&gt;
&lt;br /&gt;
===B Cell Interactome===&lt;br /&gt;
* '''Description''': The B cell interactome (BCI)  is a network of protein-protein, protein-DNA and modulatory interactions in human B cells. The network contains known interactions (reported in public databases) and predicted interactions by a Bayesian evidence integration framework which integrates a variety of generic and context specific experimental clues about protein-protein and protein-DNA interactions - such as a large collection of B cell expression profiles - with inferences from different reverse engineering algorithms, such as GeneWays and ARACNE. Modulatory interactions are predicted by MINDY, an algorithm for the prediction of modulators of transcriptional interactions.&lt;br /&gt;
* '''Data Input''': n/a&lt;br /&gt;
* '''Data Output''':  text file of binary interations associated with a probability. &lt;br /&gt;
* '''Implementation Language''': Perl&lt;br /&gt;
* '''Version, Date, Stage''': Version 2, March 2007&lt;br /&gt;
* '''Authors''':  Lefebvre C, Lim WK, Basso K, Dalla Favera R, and Califano A.&lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': Naive Bayes, Mixed-Interaction Network, human B cells.&lt;br /&gt;
* '''URL''': http://amdec-bioinfo.cu-genome.org/html/BCellInteractome.html&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Interaction Modeling&lt;br /&gt;
&lt;br /&gt;
===ARACNE===&lt;br /&gt;
* '''Description''': ARACNE is an algorithm for inferring gene regulatory networks from a set of microarray experiments. The method uses mutual information to identify genes that are co-expressed and then applies the data processing inequality to filter out interactions that are likely to be indirect.&lt;br /&gt;
* '''Data Input''': Text file containing measurements from a set of microarray experiments.&lt;br /&gt;
* '''Data Output''':  Text file containing predicted interactions.&lt;br /&gt;
* '''Implementation Language''': C++, Java&lt;br /&gt;
* '''Version, Date, Stage''': Version 1, June, 2006&lt;br /&gt;
* '''Authors''':  Margolin AA, Nemenman I, Basso K, Wiggins C, Stolovitzky G, Dalla Favera R, Califano A.&lt;br /&gt;
* '''Platforms Tested''': Window, Linux&lt;br /&gt;
* '''License''':  Open source&lt;br /&gt;
* '''Keywords''': Reverse engineering, mutual information, genetic networks, microarray&lt;br /&gt;
* '''URL''': http://amdec-bioinfo.cu-genome.org/html/ARACNE.htm&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===geWorkbench===&lt;br /&gt;
* '''Description''': geWorkbench is a Java application that provides users with an integrated suite of genomics tools. It is built on an open-source, extensible architecture that promotes interoperability and simplifies the development of new as well as the incorporation of pre-existing components. The resulting system provides seamless access to a multitude of both local and remote data and computational services through an integrated environment that offers a unified user experience. Over 50 data analysis and visualization components have been developed for the framework, covering a wide range of genomics domains including gene expression, sequence, structure and network data.&lt;br /&gt;
* '''Data Input''': Gene epxression data (Affy, GenPix, RMA), Sequence (FASTA), Structure (PDB).&lt;br /&gt;
* '''Data Output''': Analysis results (multiple formats).&lt;br /&gt;
* '''Implementation Language''': Java&lt;br /&gt;
* '''Version, Date, Stage''': 1.0.5, 3/23/07, stable production release&lt;br /&gt;
* '''Authors''':  A. Califano, A. Floratos. M. Kustagi, K. Smith, J. Watkinson, M. Hall, K. Keshav, X. Zhang, K. Kushal, B. Jagla, E. Daly, M. VanGinhoven, P. Morozov.&lt;br /&gt;
* '''Platforms Tested''': Windows XP, Linux, Mac OS 10.x.&lt;br /&gt;
* '''License''': Free. &lt;br /&gt;
* '''Keywords''': Analysis suite, gene expression analysis, sequence analysis, network reconstruction, structure predcition, visualization.&lt;br /&gt;
* '''URL''': http://www.geworkbench.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Interaction Modeling; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Software Engineering and Development Tool --&amp;gt; Integration --&amp;gt; Resource Integration Components; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Software Engineering and Development Tool --&amp;gt; Integration --&amp;gt; Grid Computing Resources; Atomic --&amp;gt; SoftwareFunction --&amp;gt; Visualization&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10270</id>
		<title>SDIWG:NCBC Software Classification MAGNet Examples</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10270"/>
		<updated>2007-05-13T21:39:01Z</updated>

		<summary type="html">&lt;p&gt;Aris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;===DelPhi===&lt;br /&gt;
* '''Description''': DelPhi provides numerical solutions to the Poisson-Boltzmann equation (both linear and nonlinear form) for molecules of arbitrary shape and charge distribution. The current version is fast (the best relaxation parameter is estimated at run time), accurate (calculation of the electrostatic free energy is less dependent on the resolution of the lattice) and can handle extremely high lattice dimensions. It also includes flexible features for assigning different dielectric constants to different regions of space and treating systems containing mixed salt solutions.&lt;br /&gt;
* '''Data Input''': DelPhi takes as input a coordinate file format of a molecule or equivalent data for geometrical objects and/or charge distributions&lt;br /&gt;
* '''Data Output''': electrostatic potential in and around the system&lt;br /&gt;
* '''Implementation Language''': Fortran and C&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''':  E.Alexov,  R.Fine, M.K.Gilson, A.Nicholls, W.Rocchia, K.Sharp, and B. Honig.&lt;br /&gt;
* '''Platforms Tested''': Unix-SGI IRIX, linux, PC (requires Fortran and C compilers), AIX IBM version and Mac.&lt;br /&gt;
* '''License''': Freely available to academia; pay model for commercial users. &lt;br /&gt;
* '''Keywords''': Finite Difference Poisson-Boltzman Solver&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/delphi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Numerical Calculation of Electrostatic Potential&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP===&lt;br /&gt;
* '''Description''': A molecular visualization and analysis program. It is particularly useful for the display and manipulation of the surfaces of molecules and their electrostatic properties.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi&lt;br /&gt;
* '''Data Output''': molecular graphics.&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': v1.3.6 .Stable public release.&lt;br /&gt;
* '''Authors''': Anthony Nicholls and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': SGI machines: irix 5.x and 6.x (INDYs, INDIGOs including Impact, Octane and O2) systems.&lt;br /&gt;
* '''License''':  Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
===Nest===&lt;br /&gt;
* '''Description''': Modeling protein structure based on a sequence-template alignment. The current server works only for modeling with a single template. Part of jackal, which can be downloaded.&lt;br /&gt;
* '''Data Input''': pir and PDB files&lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Xiang, Z. and Honig, B.&lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': modeling, protein structure, sequence-template alignment.&lt;br /&gt;
* '''URL''': http://honiglab.cpmc.columbia.edu/cgi-bin/jackal/nest.cgi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===JACKAL===&lt;br /&gt;
* '''Description''': Jackal is a collection of programs designed for the modeling and analysis of protein structures. Its core program is a versatile homology modeling package nest. JACKAL has the following capabilities: 1) comparative modeling based on single, composite or multiple templates; 2) side-chain prediction; 3) modeling residue mutation, insertion or deletion; 4) loop prediction; 5) structure refinement; 6) reconstruction of protein missing atoms;7) reconstruction of protein missing residues; 8) prediction of hydrogen atoms; 9) fast calculation of solvent accessible surface area; 10) structure superimposition.       &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Version: 1.5 as of Oct, 20, 2002, Stable public release.&lt;br /&gt;
* '''Authors''': Z. Xiang and B. Honig&lt;br /&gt;
* '''Platforms Tested''': SGI 6.5, Intel Linux and Sun solaris&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': Protein Structure Modeling&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/jackal&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP2===&lt;br /&gt;
* '''Description''': GRASP2 is an updated version of the GRASP program used for macromolecular structure and surface visualization, contains a large number of new features and scientific tools: Enhanced GUI; Structure alignment and domain database scanning; A gaussian surface generator and new surface coloring schemes; Sequence visualization and alignment; Completed work can be stored in &amp;quot;project files;  Among the many objects that can be stored in a project file are views of the structure; defined subsets, surfaces; Direct printing to printers at full printer resolution.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi, sequence alignments.&lt;br /&gt;
* '''Data Output''': molecular graphics, structural alignments.&lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Donald Petrey and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp2&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===PrISM===&lt;br /&gt;
* '''Description''': PrISM is an integrated computational system where computational tools are implemented for protein sequence and structure analysis and modeling.&lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Wang, L, Yang, A. S. &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''':SGI-irix, Intel-linux &lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': protein analysis/modeling &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/PrISM/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Protein-DNA interface alignment===&lt;br /&gt;
* '''Description''': The protein-DNA alignment software allows one to align the interfacial amino acids from two protein-DNA complexes based on the geometric relationship of each amino acid to its local DNA. &lt;br /&gt;
* '''Data Input''': two PDB files that both contain protein-DNA complexes&lt;br /&gt;
* '''Data Output''': The programs will output the aligned residues and their corresponding residue-residue similarity scores, s(i,j).&lt;br /&gt;
* '''Implementation Language''': C++ and Perl&lt;br /&gt;
* '''Version, Date, Stage''':Stable public release. &lt;br /&gt;
* '''Authors''': Siggers, T.W., Silkov, A &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''': Linux&lt;br /&gt;
* '''License''': Freely available to academia &lt;br /&gt;
* '''Keywords''': protein-DNA interface &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/intfc_aln&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
===SURFace===&lt;br /&gt;
* '''Description''': SURFace algorithms are programs that calculate solvent accessible surface area and curvature corrected solvent accessible surface area &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': &lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Nicholls, A., Sharp, K., Sridharan, S. and Honig, B.  &lt;br /&gt;
* '''Platforms Tested''': SGI&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': solvent accessible surface area  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/surf/&lt;br /&gt;
* '''Organization''': MAGNet &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Caculation of Solvent Accessible Area&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Target Explorer===&lt;br /&gt;
* '''Description''': Automated process of prediction of complex regulatory elements for specified set of transcription factors in Drosophila melanogaster genome. Target Explorer is a complex tool with user-friendly self-explanatory Web-interface that allows to user:  1. create customized library of TF binding site matrices based on user defined sets of training sequences; 2. search for new clusters of binding sites for specified set of TFs; 3.extract annotation for potential target genes.    &lt;br /&gt;
* '''Data Input''': genomic sequences&lt;br /&gt;
* '''Data Output''': clusters of known binding sites &lt;br /&gt;
* '''Implementation Language''': perl, cgi&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release. &lt;br /&gt;
* '''Authors''': Sosinsky A, Bonin CP, Mann RS, Honig B.   &lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': prediction of binding sites for transcription factors  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/Target_Explorer/&lt;br /&gt;
* '''Organization''': MAGNet  &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis--&amp;gt; Sequence Annotation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===MEDUSA and Gorgon===&lt;br /&gt;
* '''Description''': MEDUSA is an algorithm for learning predictive models of transcriptional gene regulation from gene expression and promoter sequence data.  By using a statistical learning approach based on boosing, MEDUSA learns cis regulatory motifs, condition-specific regulators, and regulatory programs that predict the differential expression of target genes.  The regulatory program is specified as an alternating decision tree (ADT).  The Java implementation of MEDUSA will allow a number of visualizations of the regulatory program and other inferred regulatory information, implemented in the accompanying Gorgon tool, including hits of significant and condition-specific motifs along the promoter sequences of target genes and regulatory network figures viewable in Cytoscape.      &lt;br /&gt;
* '''Data Input''': Discretized (up/down/baseline) gene expression data in plain text format, promoter sequences in FASTA format, list of candidate transcriptional regulators and signal transducers in plain text format.&lt;br /&gt;
* '''Data Output''': Regulatory program represented as a Java serialized object file readable by Gorgon and as a human readable XML file.  Gorgon currently generates views of learned PSSMs, positional hits along promoter sequences, and views of the ADT as HTML files, and generates network figures as Cytoscape format files.&lt;br /&gt;
* '''Implementation Language''': Java (prototyped in MATLAB) &lt;br /&gt;
* '''Version, Date, Stage''': Version 2.0, July 2006, pre-release beta version; Version 1.0 (MATLAB), April 2005, stable public release  &lt;br /&gt;
* '''Authors''': David Quigley, Manuel Middendorf, Steve Lianoglou, Anshul Kundaje, Yoav Freund, Chris Wiggins, Christina Leslie   &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux, Mac OS X&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/medusa (MATLAB),http://compbio.sytes.net:8090/medusa (Java beta version)&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===String kernel package===&lt;br /&gt;
* '''Description''':The string kernel package contains implementations for the mismatch and profile string kernels for use with support vector machine (SVM) classifiers for protein sequence classification.  Both kernels compute similarity between protein sequences based on common occurrences of k-length subsequences (&amp;quot;k-mers&amp;quot;) counted with substitutions.  Kernel functions for protein sequence data enable the training of SVMs for a range of prediction problems, in particular protein structural class prediction and remote homology detection.  A version of the Spider MATLAB machine learning package is also bundled with the code, which allows users to train SVMs and evaluate performance on test sets with the packaged software.&lt;br /&gt;
* '''Data Input''': The mismatch kernel requires sequence data in FASTA format. The profile string kernel uses probabilistic profiles, such as those produced by PSI-BLAST, in place of the original sequences.  The Spider SVM implementation requires both the kernel matrix and a label file of binary or multi-class labels for the training data; this data must be loaded into MATLAB variables before using Spider routing.&lt;br /&gt;
* '''Data Output''':The kernel code produces a kernel matrix for the input data in tab-delimited text format.  The Spider package trains SVMs and stores the learns classifier and results from applying the classifier on test data as MATLAB objects.&lt;br /&gt;
* '''Implementation Language''': String kernel code is implemented in C. Spider is a set of object-oriented MATLAB routines. &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.2, September 2004, stable public release &lt;br /&gt;
* '''Authors''': Eleazar Eskin, Rui Kuang, Eugene Ie, Ke Wang, Jason Weston, Bill Noble, Christina Leslie  &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/string-kernels&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MatrixREDUCE===&lt;br /&gt;
* '''Description''': Regulation of gene expression by a transcription factor requires physical interaction between the factor and the DNA, which can be described by astatistical mechanical model. Based on this model, the MatrixREDUCE algorithm uses genome-wide occupancy data for a transcription factor (e.g.ChIP-chip or mRNA expression data) and associated nucleotide sequences to discover the sequence-specific binding affinity of the transcription factor. The sequence specificity of the transcription factor's DNA-binding domain is modeled using a position-specific affinity matrix (PSAM), representing the change in the binding affinity (Kd) whenever a specific position within a reference binding sequence is mutated. The PSAM can be transformed into affinity logo for visualization using the utility program AffinityLogo, and a MatrixREDUCE run can be summarized in an easy-to-navigate webpage using HTMLSummary. &lt;br /&gt;
* '''Data Input''': sequence file in FASTA format; and expression data file in tab-delimited text format.&lt;br /&gt;
* '''Data Output''': PSAMs in numeric and graphical format, parameters of the fitted model, and an HTML summary page.&lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of Numerical Recipes routines.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.0, July 10, 2006, extensively tested in lab.  &lt;br /&gt;
* '''Authors''': Barrett Foat, Xiang-Jun Lu, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': position-specific affinity matrix, binding affinity, cis-regulatory element, expression data, ChIP-chip, transcription factor &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/MatrixREDUCE&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===T-profiler===&lt;br /&gt;
* '''Description''': T-profiler is a web-based tool that uses the t-test to score changes in the average activity of pre-defined groups of genes. The gene groups are defined based on Gene Ontology categorization, ChIP-chip experiments, upstream matches to a consensus transcription factor binding motif, and location on the same chromosome, respectively. If desired, an iterative procedure can be used to select a single, optimal representative from sets of overlapping gene groups. A jack-knife procedure is used to make calculations more robust against outliers. T-profiler makes it possible to interpret microarray data in a way that is both intuitive and statistically rigorous, without the need to combine experiments or choose parameters.   &lt;br /&gt;
* '''Data Input''': Currently, gene expression data from Saccharomyces cerevisiae and Candida albicans are supported. &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': T-profiler is written in PHP, data is managed by a MYSQL database server  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': André Boorsma, Barrett C. Foat, Daniel Vis, Frans Klis, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Web-based application&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': gene expression, transcriptome,  ChIP-chip,  Gene Ontology &lt;br /&gt;
* '''URL''': http://www.t-profiler.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===TranscriptionDetector===&lt;br /&gt;
* '''Description''':A tool for finding probes measuring significantly expressed loci in a genomic array experiment. Given expression data from some tiling array experiment, TranscriptionDetector decides the likelihood that a probe is detecting transcription from the locus in which it resides. Probabilities are assigned by making use of a background signal intensity distribution from a set of negative control probes. This tool is useful for the functional annotation of genomes as it allows for the discovery of novel transcriptional units independently of any genomic annotation.    &lt;br /&gt;
* '''Data Input''': Expression data (GEO or other platforms) and designation of which probes represent negative controls and which are data probes. &lt;br /&gt;
* '''Data Output''': A text file with a list of probes  corresponding to significantly expressed loci. &lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of GSL.  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': Xiang-Jun Lu, Gabor Halasz, Marinus F. van Batenburg  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': tiling arrays, expression, transcriptome &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/TranscriptionDetector/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===PhenoGO===&lt;br /&gt;
* '''Description''': PhenoGO adds phenotypic contextual information to existing associations between gene products and Gene Ontology (GO) terms as specified in GO Annotations (GOA). PhenoGO utilizes an existing Natural Language Processing (NLP) system, called BioMedLEE, an existing knowledge-based phenotype organizer system (PhenOS) in conjunction with MeSH indexing and established biomedical ontologies. The system also encodes the context to identifiers that are associated in different biomedical ontologies, including the UMLS, Cell Ontology, Mouse Anatomy, NCBI taxonomy, GO, and Mammalian Phenotype Ontology. In addition, PhenoGO was evaluated for coding of anatomical and cellular information and assigning the coded phenotypes to the correct GOA; results obtained show that PhenoGO has a precision of 91% and recall of 92%, demonstrating that the PhenoGO NLP system can accurately encode a large number of anatomical and cellular ontologies to GO annotations. The PhenoGO Database may be accessed at www.phenogo.org.    &lt;br /&gt;
* '''Data Input''': Gene Ontology Annotations Files and Medline Abstracts  &lt;br /&gt;
* '''Data Output''': XML file and www.phenogo.org Web Portal&lt;br /&gt;
* '''Implementation Language''': A variety of modules, the web portal is in Java and MySQL, the computational terminology component (phenOS) is written in Perl scripts that queries tables in IBM DB2,  the natural language processing component is written in PROLOG.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 2, Feb 2006  &lt;br /&gt;
* '''Authors''': Yves Lussier and Carol Friedman are the principal investigators. The programmers are Jianrong Li, Lee Sam, and Tara Borlawsky   &lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': Phenotypic integration, computational phenotypes &lt;br /&gt;
* '''URL''': www.phenogo.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MINDY===&lt;br /&gt;
* '''Description''': Given a transcription factor of interest, MINDY uses a large set of gene expression profile data to identify potential post-transcriptional modulators of the transcription factor's activity. MINDY is based on a three-way statistical interaction model that captures the post-transcriptional regulatory event where the ability of a transcription factor to activate/repress its target genes is monotonically controlled by a potential modulator gene. &lt;br /&gt;
* '''Data Input''': Gene expression data in the EXP format, and a user-specified transcription factor of interest&lt;br /&gt;
* '''Data Output''': Lists of the putative modulators and target genes of the transcription factor, and the modulatory interactions involving them&lt;br /&gt;
* '''Implementation Language''': C++ and MATLAB, Java&lt;br /&gt;
* '''Version, Date, Stage''': Stable release, April 2007  &lt;br /&gt;
* '''Authors''': Kai Wang, Ilya Nemenman, Adam Margolin, Riccardo Dalla-Favera, Andrea Califano&lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': gene expression, transcriptional interaction, modulator &lt;br /&gt;
* '''URL''': n/a&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;br /&gt;
&lt;br /&gt;
===B Cell Interactome===&lt;br /&gt;
* '''Description''': The B cell interactome (BCI)  is a network of protein-protein, protein-DNA and modulatory interactions in human B cells. The network contains known interactions (reported in public databases) and predicted interactions by a Bayesian evidence integration framework which integrates a variety of generic and context specific experimental clues about protein-protein and protein-DNA interactions - such as a large collection of B cell expression profiles - with inferences from different reverse engineering algorithms, such as GeneWays and ARACNE. Modulatory interactions are predicted by MINDY, an algorithm for the prediction of modulators of transcriptional interactions.&lt;br /&gt;
* '''Data Input''': n/a&lt;br /&gt;
* '''Data Output''':  text file of binary interations associated with a probability. &lt;br /&gt;
* '''Implementation Language''': Perl&lt;br /&gt;
* '''Version, Date, Stage''': Version 2, March 2007&lt;br /&gt;
* '''Authors''':  Lefebvre C, Lim WK, Basso K, Dalla Favera R, and Califano A.&lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': Naive Bayes, Mixed-Interaction Network, human B cells.&lt;br /&gt;
* '''URL''': http://amdec-bioinfo.cu-genome.org/html/BCellInteractome.html&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Interaction Modeling&lt;br /&gt;
&lt;br /&gt;
===ARACNE===&lt;br /&gt;
* '''Description''': ARACNE is an algorithm for inferring gene regulatory networks from a set of microarray experiments. The method uses mutual information to identify genes that are co-expressed and then applies the data processing inequality to filter out interactions that are likely to be indirect.&lt;br /&gt;
* '''Data Input''': Text file containing measurements from a set of microarray experiments.&lt;br /&gt;
* '''Data Output''':  Text file containing predicted interactions.&lt;br /&gt;
* '''Implementation Language''': C++, Java&lt;br /&gt;
* '''Version, Date, Stage''': Version 1, June, 2006&lt;br /&gt;
* '''Authors''':  Margolin AA, Nemenman I, Basso K, Wiggins C, Stolovitzky G, Dalla Favera R, Califano A.&lt;br /&gt;
* '''Platforms Tested''': Window, Linux&lt;br /&gt;
* '''License''':  Open source&lt;br /&gt;
* '''Keywords''': Reverse engineering, mutual information, genetic networks, microarray&lt;br /&gt;
* '''URL''': http://amdec-bioinfo.cu-genome.org/html/ARACNE.htm&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===geWorkbench===&lt;br /&gt;
* '''Description''': geWorkbench is a Java application that provides users with an integrated suite of genomics tools. It is built on an open-source, extensible architecture that promotes interoperability and simplifies the development of new as well as the incorporation of pre-existing components. The resulting system provides seamless access to a multitude of both local and remote data and computational services through an integrated environment that offers a unified user experience. Over 50 data analysis and visualization components have been developed for the framework, covering a wide range of genomics domains including gene expression, sequence, structure and network data.&lt;br /&gt;
* '''Data Input''': Gene epxression data (Affy, GenPix, RMA), Sequence (FASTA), Structure (PDB).&lt;br /&gt;
* '''Data Output''': Analysis results (multiple formats).&lt;br /&gt;
* '''Implementation Language''': Java&lt;br /&gt;
* '''Version, Date, Stage''': 1.0.5, 3/23/07, stable production release&lt;br /&gt;
* '''Authors''':  A. Califano, A. Floratos. M. Kustagi, K. Smith, J. Watkinson, M. Hall, K. Keshav, X. Zhang, K. Kushal, B. Jagla, E. Daly, M. VanGinhoven, P. Morozov.&lt;br /&gt;
* '''Platforms Tested''': Windows XP, Linux, Mac OS 10.x.&lt;br /&gt;
* '''License''': Free. &lt;br /&gt;
* '''Keywords''': Analysis suite, gene expression analysis, sequence analysis, network reconstruction, structure predcition, visualization.&lt;br /&gt;
* '''URL''': http://www.geworkbench.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis, Atomic --&amp;gt; SoftwareFunction --&amp;gt; Interaction Modeling, Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification, Atomic --&amp;gt; SoftwareFunction --&amp;gt; Software Engineering and Development Tool --&amp;gt; Integration --&amp;gt; Resource Integration Components, Atomic --&amp;gt; SoftwareFunction --&amp;gt; Software Engineering and Development Tool --&amp;gt; Integration --&amp;gt; Grid Computing Resources, Atomic --&amp;gt; SoftwareFunction --&amp;gt; Visualization&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10269</id>
		<title>SDIWG:NCBC Software Classification MAGNet Examples</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10269"/>
		<updated>2007-05-13T21:18:30Z</updated>

		<summary type="html">&lt;p&gt;Aris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;===DelPhi===&lt;br /&gt;
* '''Description''': DelPhi provides numerical solutions to the Poisson-Boltzmann equation (both linear and nonlinear form) for molecules of arbitrary shape and charge distribution. The current version is fast (the best relaxation parameter is estimated at run time), accurate (calculation of the electrostatic free energy is less dependent on the resolution of the lattice) and can handle extremely high lattice dimensions. It also includes flexible features for assigning different dielectric constants to different regions of space and treating systems containing mixed salt solutions.&lt;br /&gt;
* '''Data Input''': DelPhi takes as input a coordinate file format of a molecule or equivalent data for geometrical objects and/or charge distributions&lt;br /&gt;
* '''Data Output''': electrostatic potential in and around the system&lt;br /&gt;
* '''Implementation Language''': Fortran and C&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''':  E.Alexov,  R.Fine, M.K.Gilson, A.Nicholls, W.Rocchia, K.Sharp, and B. Honig.&lt;br /&gt;
* '''Platforms Tested''': Unix-SGI IRIX, linux, PC (requires Fortran and C compilers), AIX IBM version and Mac.&lt;br /&gt;
* '''License''': Freely available to academia; pay model for commercial users. &lt;br /&gt;
* '''Keywords''': Finite Difference Poisson-Boltzman Solver&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/delphi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Numerical Calculation of Electrostatic Potential&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP===&lt;br /&gt;
* '''Description''': A molecular visualization and analysis program. It is particularly useful for the display and manipulation of the surfaces of molecules and their electrostatic properties.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi&lt;br /&gt;
* '''Data Output''': molecular graphics.&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': v1.3.6 .Stable public release.&lt;br /&gt;
* '''Authors''': Anthony Nicholls and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': SGI machines: irix 5.x and 6.x (INDYs, INDIGOs including Impact, Octane and O2) systems.&lt;br /&gt;
* '''License''':  Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
===Nest===&lt;br /&gt;
* '''Description''': Modeling protein structure based on a sequence-template alignment. The current server works only for modeling with a single template. Part of jackal, which can be downloaded.&lt;br /&gt;
* '''Data Input''': pir and PDB files&lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Xiang, Z. and Honig, B.&lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': modeling, protein structure, sequence-template alignment.&lt;br /&gt;
* '''URL''': http://honiglab.cpmc.columbia.edu/cgi-bin/jackal/nest.cgi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===JACKAL===&lt;br /&gt;
* '''Description''': Jackal is a collection of programs designed for the modeling and analysis of protein structures. Its core program is a versatile homology modeling package nest. JACKAL has the following capabilities: 1) comparative modeling based on single, composite or multiple templates; 2) side-chain prediction; 3) modeling residue mutation, insertion or deletion; 4) loop prediction; 5) structure refinement; 6) reconstruction of protein missing atoms;7) reconstruction of protein missing residues; 8) prediction of hydrogen atoms; 9) fast calculation of solvent accessible surface area; 10) structure superimposition.       &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Version: 1.5 as of Oct, 20, 2002, Stable public release.&lt;br /&gt;
* '''Authors''': Z. Xiang and B. Honig&lt;br /&gt;
* '''Platforms Tested''': SGI 6.5, Intel Linux and Sun solaris&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': Protein Structure Modeling&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/jackal&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP2===&lt;br /&gt;
* '''Description''': GRASP2 is an updated version of the GRASP program used for macromolecular structure and surface visualization, contains a large number of new features and scientific tools: Enhanced GUI; Structure alignment and domain database scanning; A gaussian surface generator and new surface coloring schemes; Sequence visualization and alignment; Completed work can be stored in &amp;quot;project files;  Among the many objects that can be stored in a project file are views of the structure; defined subsets, surfaces; Direct printing to printers at full printer resolution.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi, sequence alignments.&lt;br /&gt;
* '''Data Output''': molecular graphics, structural alignments.&lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Donald Petrey and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp2&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===PrISM===&lt;br /&gt;
* '''Description''': PrISM is an integrated computational system where computational tools are implemented for protein sequence and structure analysis and modeling.&lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Wang, L, Yang, A. S. &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''':SGI-irix, Intel-linux &lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': protein analysis/modeling &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/PrISM/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Protein-DNA interface alignment===&lt;br /&gt;
* '''Description''': The protein-DNA alignment software allows one to align the interfacial amino acids from two protein-DNA complexes based on the geometric relationship of each amino acid to its local DNA. &lt;br /&gt;
* '''Data Input''': two PDB files that both contain protein-DNA complexes&lt;br /&gt;
* '''Data Output''': The programs will output the aligned residues and their corresponding residue-residue similarity scores, s(i,j).&lt;br /&gt;
* '''Implementation Language''': C++ and Perl&lt;br /&gt;
* '''Version, Date, Stage''':Stable public release. &lt;br /&gt;
* '''Authors''': Siggers, T.W., Silkov, A &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''': Linux&lt;br /&gt;
* '''License''': Freely available to academia &lt;br /&gt;
* '''Keywords''': protein-DNA interface &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/intfc_aln&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
===SURFace===&lt;br /&gt;
* '''Description''': SURFace algorithms are programs that calculate solvent accessible surface area and curvature corrected solvent accessible surface area &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': &lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Nicholls, A., Sharp, K., Sridharan, S. and Honig, B.  &lt;br /&gt;
* '''Platforms Tested''': SGI&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': solvent accessible surface area  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/surf/&lt;br /&gt;
* '''Organization''': MAGNet &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Caculation of Solvent Accessible Area&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Target Explorer===&lt;br /&gt;
* '''Description''': Automated process of prediction of complex regulatory elements for specified set of transcription factors in Drosophila melanogaster genome. Target Explorer is a complex tool with user-friendly self-explanatory Web-interface that allows to user:  1. create customized library of TF binding site matrices based on user defined sets of training sequences; 2. search for new clusters of binding sites for specified set of TFs; 3.extract annotation for potential target genes.    &lt;br /&gt;
* '''Data Input''': genomic sequences&lt;br /&gt;
* '''Data Output''': clusters of known binding sites &lt;br /&gt;
* '''Implementation Language''': perl, cgi&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release. &lt;br /&gt;
* '''Authors''': Sosinsky A, Bonin CP, Mann RS, Honig B.   &lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': prediction of binding sites for transcription factors  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/Target_Explorer/&lt;br /&gt;
* '''Organization''': MAGNet  &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis--&amp;gt; Sequence Annotation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===MEDUSA and Gorgon===&lt;br /&gt;
* '''Description''': MEDUSA is an algorithm for learning predictive models of transcriptional gene regulation from gene expression and promoter sequence data.  By using a statistical learning approach based on boosing, MEDUSA learns cis regulatory motifs, condition-specific regulators, and regulatory programs that predict the differential expression of target genes.  The regulatory program is specified as an alternating decision tree (ADT).  The Java implementation of MEDUSA will allow a number of visualizations of the regulatory program and other inferred regulatory information, implemented in the accompanying Gorgon tool, including hits of significant and condition-specific motifs along the promoter sequences of target genes and regulatory network figures viewable in Cytoscape.      &lt;br /&gt;
* '''Data Input''': Discretized (up/down/baseline) gene expression data in plain text format, promoter sequences in FASTA format, list of candidate transcriptional regulators and signal transducers in plain text format.&lt;br /&gt;
* '''Data Output''': Regulatory program represented as a Java serialized object file readable by Gorgon and as a human readable XML file.  Gorgon currently generates views of learned PSSMs, positional hits along promoter sequences, and views of the ADT as HTML files, and generates network figures as Cytoscape format files.&lt;br /&gt;
* '''Implementation Language''': Java (prototyped in MATLAB) &lt;br /&gt;
* '''Version, Date, Stage''': Version 2.0, July 2006, pre-release beta version; Version 1.0 (MATLAB), April 2005, stable public release  &lt;br /&gt;
* '''Authors''': David Quigley, Manuel Middendorf, Steve Lianoglou, Anshul Kundaje, Yoav Freund, Chris Wiggins, Christina Leslie   &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux, Mac OS X&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/medusa (MATLAB),http://compbio.sytes.net:8090/medusa (Java beta version)&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===String kernel package===&lt;br /&gt;
* '''Description''':The string kernel package contains implementations for the mismatch and profile string kernels for use with support vector machine (SVM) classifiers for protein sequence classification.  Both kernels compute similarity between protein sequences based on common occurrences of k-length subsequences (&amp;quot;k-mers&amp;quot;) counted with substitutions.  Kernel functions for protein sequence data enable the training of SVMs for a range of prediction problems, in particular protein structural class prediction and remote homology detection.  A version of the Spider MATLAB machine learning package is also bundled with the code, which allows users to train SVMs and evaluate performance on test sets with the packaged software.&lt;br /&gt;
* '''Data Input''': The mismatch kernel requires sequence data in FASTA format. The profile string kernel uses probabilistic profiles, such as those produced by PSI-BLAST, in place of the original sequences.  The Spider SVM implementation requires both the kernel matrix and a label file of binary or multi-class labels for the training data; this data must be loaded into MATLAB variables before using Spider routing.&lt;br /&gt;
* '''Data Output''':The kernel code produces a kernel matrix for the input data in tab-delimited text format.  The Spider package trains SVMs and stores the learns classifier and results from applying the classifier on test data as MATLAB objects.&lt;br /&gt;
* '''Implementation Language''': String kernel code is implemented in C. Spider is a set of object-oriented MATLAB routines. &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.2, September 2004, stable public release &lt;br /&gt;
* '''Authors''': Eleazar Eskin, Rui Kuang, Eugene Ie, Ke Wang, Jason Weston, Bill Noble, Christina Leslie  &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/string-kernels&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MatrixREDUCE===&lt;br /&gt;
* '''Description''': Regulation of gene expression by a transcription factor requires physical interaction between the factor and the DNA, which can be described by astatistical mechanical model. Based on this model, the MatrixREDUCE algorithm uses genome-wide occupancy data for a transcription factor (e.g.ChIP-chip or mRNA expression data) and associated nucleotide sequences to discover the sequence-specific binding affinity of the transcription factor. The sequence specificity of the transcription factor's DNA-binding domain is modeled using a position-specific affinity matrix (PSAM), representing the change in the binding affinity (Kd) whenever a specific position within a reference binding sequence is mutated. The PSAM can be transformed into affinity logo for visualization using the utility program AffinityLogo, and a MatrixREDUCE run can be summarized in an easy-to-navigate webpage using HTMLSummary. &lt;br /&gt;
* '''Data Input''': sequence file in FASTA format; and expression data file in tab-delimited text format.&lt;br /&gt;
* '''Data Output''': PSAMs in numeric and graphical format, parameters of the fitted model, and an HTML summary page.&lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of Numerical Recipes routines.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.0, July 10, 2006, extensively tested in lab.  &lt;br /&gt;
* '''Authors''': Barrett Foat, Xiang-Jun Lu, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': position-specific affinity matrix, binding affinity, cis-regulatory element, expression data, ChIP-chip, transcription factor &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/MatrixREDUCE&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===T-profiler===&lt;br /&gt;
* '''Description''': T-profiler is a web-based tool that uses the t-test to score changes in the average activity of pre-defined groups of genes. The gene groups are defined based on Gene Ontology categorization, ChIP-chip experiments, upstream matches to a consensus transcription factor binding motif, and location on the same chromosome, respectively. If desired, an iterative procedure can be used to select a single, optimal representative from sets of overlapping gene groups. A jack-knife procedure is used to make calculations more robust against outliers. T-profiler makes it possible to interpret microarray data in a way that is both intuitive and statistically rigorous, without the need to combine experiments or choose parameters.   &lt;br /&gt;
* '''Data Input''': Currently, gene expression data from Saccharomyces cerevisiae and Candida albicans are supported. &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': T-profiler is written in PHP, data is managed by a MYSQL database server  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': André Boorsma, Barrett C. Foat, Daniel Vis, Frans Klis, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Web-based application&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': gene expression, transcriptome,  ChIP-chip,  Gene Ontology &lt;br /&gt;
* '''URL''': http://www.t-profiler.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===TranscriptionDetector===&lt;br /&gt;
* '''Description''':A tool for finding probes measuring significantly expressed loci in a genomic array experiment. Given expression data from some tiling array experiment, TranscriptionDetector decides the likelihood that a probe is detecting transcription from the locus in which it resides. Probabilities are assigned by making use of a background signal intensity distribution from a set of negative control probes. This tool is useful for the functional annotation of genomes as it allows for the discovery of novel transcriptional units independently of any genomic annotation.    &lt;br /&gt;
* '''Data Input''': Expression data (GEO or other platforms) and designation of which probes represent negative controls and which are data probes. &lt;br /&gt;
* '''Data Output''': A text file with a list of probes  corresponding to significantly expressed loci. &lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of GSL.  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': Xiang-Jun Lu, Gabor Halasz, Marinus F. van Batenburg  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': tiling arrays, expression, transcriptome &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/TranscriptionDetector/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===PhenoGO===&lt;br /&gt;
* '''Description''': PhenoGO adds phenotypic contextual information to existing associations between gene products and Gene Ontology (GO) terms as specified in GO Annotations (GOA). PhenoGO utilizes an existing Natural Language Processing (NLP) system, called BioMedLEE, an existing knowledge-based phenotype organizer system (PhenOS) in conjunction with MeSH indexing and established biomedical ontologies. The system also encodes the context to identifiers that are associated in different biomedical ontologies, including the UMLS, Cell Ontology, Mouse Anatomy, NCBI taxonomy, GO, and Mammalian Phenotype Ontology. In addition, PhenoGO was evaluated for coding of anatomical and cellular information and assigning the coded phenotypes to the correct GOA; results obtained show that PhenoGO has a precision of 91% and recall of 92%, demonstrating that the PhenoGO NLP system can accurately encode a large number of anatomical and cellular ontologies to GO annotations. The PhenoGO Database may be accessed at www.phenogo.org.    &lt;br /&gt;
* '''Data Input''': Gene Ontology Annotations Files and Medline Abstracts  &lt;br /&gt;
* '''Data Output''': XML file and www.phenogo.org Web Portal&lt;br /&gt;
* '''Implementation Language''': A variety of modules, the web portal is in Java and MySQL, the computational terminology component (phenOS) is written in Perl scripts that queries tables in IBM DB2,  the natural language processing component is written in PROLOG.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 2, Feb 2006  &lt;br /&gt;
* '''Authors''': Yves Lussier and Carol Friedman are the principal investigators. The programmers are Jianrong Li, Lee Sam, and Tara Borlawsky   &lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': Phenotypic integration, computational phenotypes &lt;br /&gt;
* '''URL''': www.phenogo.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MINDY===&lt;br /&gt;
* '''Description''': Given a transcription factor of interest, MINDY uses a large set of gene expression profile data to identify potential post-transcriptional modulators of the transcription factor's activity. MINDY is based on a three-way statistical interaction model that captures the post-transcriptional regulatory event where the ability of a transcription factor to activate/repress its target genes is monotonically controlled by a potential modulator gene. &lt;br /&gt;
* '''Data Input''': Gene expression data in the EXP format, and a user-specified transcription factor of interest&lt;br /&gt;
* '''Data Output''': Lists of the putative modulators and target genes of the transcription factor, and the modulatory interactions involving them&lt;br /&gt;
* '''Implementation Language''': C++ and MATLAB, Java&lt;br /&gt;
* '''Version, Date, Stage''': Stable release, April 2007  &lt;br /&gt;
* '''Authors''': Kai Wang, Ilya Nemenman, Adam Margolin, Riccardo Dalla-Favera, Andrea Califano&lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': gene expression, transcriptional interaction, modulator &lt;br /&gt;
* '''URL''': n/a&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;br /&gt;
&lt;br /&gt;
===B Cell Interactome===&lt;br /&gt;
* '''Description''': The B cell interactome (BCI)  is a network of protein-protein, protein-DNA and modulatory interactions in human B cells. The network contains known interactions (reported in public databases) and predicted interactions by a Bayesian evidence integration framework which integrates a variety of generic and context specific experimental clues about protein-protein and protein-DNA interactions - such as a large collection of B cell expression profiles - with inferences from different reverse engineering algorithms, such as GeneWays and ARACNE. Modulatory interactions are predicted by MINDY, an algorithm for the prediction of modulators of transcriptional interactions.&lt;br /&gt;
* '''Data Input''': n/a&lt;br /&gt;
* '''Data Output''':  text file of binary interations associated with a probability. &lt;br /&gt;
* '''Implementation Language''': Perl&lt;br /&gt;
* '''Version, Date, Stage''': Version 2, March 2007&lt;br /&gt;
* '''Authors''':  Lefebvre C, Lim WK, Basso K, Dalla Favera R, and Califano A.&lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': Naive Bayes, Mixed-Interaction Network, human B cells.&lt;br /&gt;
* '''URL''': http://amdec-bioinfo.cu-genome.org/html/BCellInteractome.html&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Interaction Modeling&lt;br /&gt;
&lt;br /&gt;
===ARACNE===&lt;br /&gt;
* '''Description''': ARACNE is an algorithm for inferring gene regulatory networks from a set of microarray experiments. The method uses mutual information to identify genes that are co-expressed and then applies the data processing inequality to filter out interactions that are likely to be indirect.&lt;br /&gt;
* '''Data Input''': Text file containing measurements from a set of microarray experiments.&lt;br /&gt;
* '''Data Output''':  Text file containing predicted interactions.&lt;br /&gt;
* '''Implementation Language''': C++, Java&lt;br /&gt;
* '''Version, Date, Stage''': Version 1, June, 2006&lt;br /&gt;
* '''Authors''':  Margolin AA, Nemenman I, Basso K, Wiggins C, Stolovitzky G, Dalla Favera R, Califano A.&lt;br /&gt;
* '''Platforms Tested''': Window, Linux&lt;br /&gt;
* '''License''':  Open source&lt;br /&gt;
* '''Keywords''': Reverse engineering, mutual information, genetic networks, microarray&lt;br /&gt;
* '''URL''': http://amdec-bioinfo.cu-genome.org/html/ARACNE.htm&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10252</id>
		<title>SDIWG:NCBC Software Classification MAGNet Examples</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10252"/>
		<updated>2007-05-11T22:26:52Z</updated>

		<summary type="html">&lt;p&gt;Aris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;===DelPhi===&lt;br /&gt;
* '''Description''': DelPhi provides numerical solutions to the Poisson-Boltzmann equation (both linear and nonlinear form) for molecules of arbitrary shape and charge distribution. The current version is fast (the best relaxation parameter is estimated at run time), accurate (calculation of the electrostatic free energy is less dependent on the resolution of the lattice) and can handle extremely high lattice dimensions. It also includes flexible features for assigning different dielectric constants to different regions of space and treating systems containing mixed salt solutions.&lt;br /&gt;
* '''Data Input''': DelPhi takes as input a coordinate file format of a molecule or equivalent data for geometrical objects and/or charge distributions&lt;br /&gt;
* '''Data Output''': electrostatic potential in and around the system&lt;br /&gt;
* '''Implementation Language''': Fortran and C&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''':  E.Alexov,  R.Fine, M.K.Gilson, A.Nicholls, W.Rocchia, K.Sharp, and B. Honig.&lt;br /&gt;
* '''Platforms Tested''': Unix-SGI IRIX, linux, PC (requires Fortran and C compilers), AIX IBM version and Mac.&lt;br /&gt;
* '''License''': Freely available to academia; pay model for commercial users. &lt;br /&gt;
* '''Keywords''': Finite Difference Poisson-Boltzman Solver&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/delphi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Numerical Calculation of Electrostatic Potential&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP===&lt;br /&gt;
* '''Description''': A molecular visualization and analysis program. It is particularly useful for the display and manipulation of the surfaces of molecules and their electrostatic properties.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi&lt;br /&gt;
* '''Data Output''': molecular graphics.&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': v1.3.6 .Stable public release.&lt;br /&gt;
* '''Authors''': Anthony Nicholls and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': SGI machines: irix 5.x and 6.x (INDYs, INDIGOs including Impact, Octane and O2) systems.&lt;br /&gt;
* '''License''':  Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
===Nest===&lt;br /&gt;
* '''Description''': Modeling protein structure based on a sequence-template alignment. The current server works only for modeling with a single template. Part of jackal, which can be downloaded.&lt;br /&gt;
* '''Data Input''': pir and PDB files&lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Xiang, Z. and Honig, B.&lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': modeling, protein structure, sequence-template alignment.&lt;br /&gt;
* '''URL''': http://honiglab.cpmc.columbia.edu/cgi-bin/jackal/nest.cgi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===JACKAL===&lt;br /&gt;
* '''Description''': Jackal is a collection of programs designed for the modeling and analysis of protein structures. Its core program is a versatile homology modeling package nest. JACKAL has the following capabilities: 1) comparative modeling based on single, composite or multiple templates; 2) side-chain prediction; 3) modeling residue mutation, insertion or deletion; 4) loop prediction; 5) structure refinement; 6) reconstruction of protein missing atoms;7) reconstruction of protein missing residues; 8) prediction of hydrogen atoms; 9) fast calculation of solvent accessible surface area; 10) structure superimposition.       &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Version: 1.5 as of Oct, 20, 2002, Stable public release.&lt;br /&gt;
* '''Authors''': Z. Xiang and B. Honig&lt;br /&gt;
* '''Platforms Tested''': SGI 6.5, Intel Linux and Sun solaris&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': Protein Structure Modeling&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/jackal&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP2===&lt;br /&gt;
* '''Description''': GRASP2 is an updated version of the GRASP program used for macromolecular structure and surface visualization, contains a large number of new features and scientific tools: Enhanced GUI; Structure alignment and domain database scanning; A gaussian surface generator and new surface coloring schemes; Sequence visualization and alignment; Completed work can be stored in &amp;quot;project files;  Among the many objects that can be stored in a project file are views of the structure; defined subsets, surfaces; Direct printing to printers at full printer resolution.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi, sequence alignments.&lt;br /&gt;
* '''Data Output''': molecular graphics, structural alignments.&lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Donald Petrey and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp2&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===PrISM===&lt;br /&gt;
* '''Description''': PrISM is an integrated computational system where computational tools are implemented for protein sequence and structure analysis and modeling.&lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Wang, L, Yang, A. S. &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''':SGI-irix, Intel-linux &lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': protein analysis/modeling &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/PrISM/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Protein-DNA interface alignment===&lt;br /&gt;
* '''Description''': The protein-DNA alignment software allows one to align the interfacial amino acids from two protein-DNA complexes based on the geometric relationship of each amino acid to its local DNA. &lt;br /&gt;
* '''Data Input''': two PDB files that both contain protein-DNA complexes&lt;br /&gt;
* '''Data Output''': The programs will output the aligned residues and their corresponding residue-residue similarity scores, s(i,j).&lt;br /&gt;
* '''Implementation Language''': C++ and Perl&lt;br /&gt;
* '''Version, Date, Stage''':Stable public release. &lt;br /&gt;
* '''Authors''': Siggers, T.W., Silkov, A &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''': Linux&lt;br /&gt;
* '''License''': Freely available to academia &lt;br /&gt;
* '''Keywords''': protein-DNA interface &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/intfc_aln&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
===SURFace===&lt;br /&gt;
* '''Description''': SURFace algorithms are programs that calculate solvent accessible surface area and curvature corrected solvent accessible surface area &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': &lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Nicholls, A., Sharp, K., Sridharan, S. and Honig, B.  &lt;br /&gt;
* '''Platforms Tested''': SGI&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': solvent accessible surface area  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/surf/&lt;br /&gt;
* '''Organization''': MAGNet &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Caculation of Solvent Accessible Area&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Target Explorer===&lt;br /&gt;
* '''Description''': Automated process of prediction of complex regulatory elements for specified set of transcription factors in Drosophila melanogaster genome. Target Explorer is a complex tool with user-friendly self-explanatory Web-interface that allows to user:  1. create customized library of TF binding site matrices based on user defined sets of training sequences; 2. search for new clusters of binding sites for specified set of TFs; 3.extract annotation for potential target genes.    &lt;br /&gt;
* '''Data Input''': genomic sequences&lt;br /&gt;
* '''Data Output''': clusters of known binding sites &lt;br /&gt;
* '''Implementation Language''': perl, cgi&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release. &lt;br /&gt;
* '''Authors''': Sosinsky A, Bonin CP, Mann RS, Honig B.   &lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': prediction of binding sites for transcription factors  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/Target_Explorer/&lt;br /&gt;
* '''Organization''': MAGNet  &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis--&amp;gt; Sequence Annotation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===MEDUSA and Gorgon===&lt;br /&gt;
* '''Description''': MEDUSA is an algorithm for learning predictive models of transcriptional gene regulation from gene expression and promoter sequence data.  By using a statistical learning approach based on boosing, MEDUSA learns cis regulatory motifs, condition-specific regulators, and regulatory programs that predict the differential expression of target genes.  The regulatory program is specified as an alternating decision tree (ADT).  The Java implementation of MEDUSA will allow a number of visualizations of the regulatory program and other inferred regulatory information, implemented in the accompanying Gorgon tool, including hits of significant and condition-specific motifs along the promoter sequences of target genes and regulatory network figures viewable in Cytoscape.      &lt;br /&gt;
* '''Data Input''': Discretized (up/down/baseline) gene expression data in plain text format, promoter sequences in FASTA format, list of candidate transcriptional regulators and signal transducers in plain text format.&lt;br /&gt;
* '''Data Output''': Regulatory program represented as a Java serialized object file readable by Gorgon and as a human readable XML file.  Gorgon currently generates views of learned PSSMs, positional hits along promoter sequences, and views of the ADT as HTML files, and generates network figures as Cytoscape format files.&lt;br /&gt;
* '''Implementation Language''': Java (prototyped in MATLAB) &lt;br /&gt;
* '''Version, Date, Stage''': Version 2.0, July 2006, pre-release beta version; Version 1.0 (MATLAB), April 2005, stable public release  &lt;br /&gt;
* '''Authors''': David Quigley, Manuel Middendorf, Steve Lianoglou, Anshul Kundaje, Yoav Freund, Chris Wiggins, Christina Leslie   &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux, Mac OS X&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/medusa (MATLAB),http://compbio.sytes.net:8090/medusa (Java beta version)&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===String kernel package===&lt;br /&gt;
* '''Description''':The string kernel package contains implementations for the mismatch and profile string kernels for use with support vector machine (SVM) classifiers for protein sequence classification.  Both kernels compute similarity between protein sequences based on common occurrences of k-length subsequences (&amp;quot;k-mers&amp;quot;) counted with substitutions.  Kernel functions for protein sequence data enable the training of SVMs for a range of prediction problems, in particular protein structural class prediction and remote homology detection.  A version of the Spider MATLAB machine learning package is also bundled with the code, which allows users to train SVMs and evaluate performance on test sets with the packaged software.&lt;br /&gt;
* '''Data Input''': The mismatch kernel requires sequence data in FASTA format. The profile string kernel uses probabilistic profiles, such as those produced by PSI-BLAST, in place of the original sequences.  The Spider SVM implementation requires both the kernel matrix and a label file of binary or multi-class labels for the training data; this data must be loaded into MATLAB variables before using Spider routing.&lt;br /&gt;
* '''Data Output''':The kernel code produces a kernel matrix for the input data in tab-delimited text format.  The Spider package trains SVMs and stores the learns classifier and results from applying the classifier on test data as MATLAB objects.&lt;br /&gt;
* '''Implementation Language''': String kernel code is implemented in C. Spider is a set of object-oriented MATLAB routines. &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.2, September 2004, stable public release &lt;br /&gt;
* '''Authors''': Eleazar Eskin, Rui Kuang, Eugene Ie, Ke Wang, Jason Weston, Bill Noble, Christina Leslie  &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/string-kernels&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MatrixREDUCE===&lt;br /&gt;
* '''Description''': Regulation of gene expression by a transcription factor requires physical interaction between the factor and the DNA, which can be described by astatistical mechanical model. Based on this model, the MatrixREDUCE algorithm uses genome-wide occupancy data for a transcription factor (e.g.ChIP-chip or mRNA expression data) and associated nucleotide sequences to discover the sequence-specific binding affinity of the transcription factor. The sequence specificity of the transcription factor's DNA-binding domain is modeled using a position-specific affinity matrix (PSAM), representing the change in the binding affinity (Kd) whenever a specific position within a reference binding sequence is mutated. The PSAM can be transformed into affinity logo for visualization using the utility program AffinityLogo, and a MatrixREDUCE run can be summarized in an easy-to-navigate webpage using HTMLSummary. &lt;br /&gt;
* '''Data Input''': sequence file in FASTA format; and expression data file in tab-delimited text format.&lt;br /&gt;
* '''Data Output''': PSAMs in numeric and graphical format, parameters of the fitted model, and an HTML summary page.&lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of Numerical Recipes routines.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.0, July 10, 2006, extensively tested in lab.  &lt;br /&gt;
* '''Authors''': Barrett Foat, Xiang-Jun Lu, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': position-specific affinity matrix, binding affinity, cis-regulatory element, expression data, ChIP-chip, transcription factor &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/MatrixREDUCE&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===T-profiler===&lt;br /&gt;
* '''Description''': T-profiler is a web-based tool that uses the t-test to score changes in the average activity of pre-defined groups of genes. The gene groups are defined based on Gene Ontology categorization, ChIP-chip experiments, upstream matches to a consensus transcription factor binding motif, and location on the same chromosome, respectively. If desired, an iterative procedure can be used to select a single, optimal representative from sets of overlapping gene groups. A jack-knife procedure is used to make calculations more robust against outliers. T-profiler makes it possible to interpret microarray data in a way that is both intuitive and statistically rigorous, without the need to combine experiments or choose parameters.   &lt;br /&gt;
* '''Data Input''': Currently, gene expression data from Saccharomyces cerevisiae and Candida albicans are supported. &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': T-profiler is written in PHP, data is managed by a MYSQL database server  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': André Boorsma, Barrett C. Foat, Daniel Vis, Frans Klis, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Web-based application&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': gene expression, transcriptome,  ChIP-chip,  Gene Ontology &lt;br /&gt;
* '''URL''': http://www.t-profiler.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===TranscriptionDetector===&lt;br /&gt;
* '''Description''':A tool for finding probes measuring significantly expressed loci in a genomic array experiment. Given expression data from some tiling array experiment, TranscriptionDetector decides the likelihood that a probe is detecting transcription from the locus in which it resides. Probabilities are assigned by making use of a background signal intensity distribution from a set of negative control probes. This tool is useful for the functional annotation of genomes as it allows for the discovery of novel transcriptional units independently of any genomic annotation.    &lt;br /&gt;
* '''Data Input''': Expression data (GEO or other platforms) and designation of which probes represent negative controls and which are data probes. &lt;br /&gt;
* '''Data Output''': A text file with a list of probes  corresponding to significantly expressed loci. &lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of GSL.  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': Xiang-Jun Lu, Gabor Halasz, Marinus F. van Batenburg  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': tiling arrays, expression, transcriptome &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/TranscriptionDetector/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===PhenoGO===&lt;br /&gt;
* '''Description''': PhenoGO adds phenotypic contextual information to existing associations between gene products and Gene Ontology (GO) terms as specified in GO Annotations (GOA). PhenoGO utilizes an existing Natural Language Processing (NLP) system, called BioMedLEE, an existing knowledge-based phenotype organizer system (PhenOS) in conjunction with MeSH indexing and established biomedical ontologies. The system also encodes the context to identifiers that are associated in different biomedical ontologies, including the UMLS, Cell Ontology, Mouse Anatomy, NCBI taxonomy, GO, and Mammalian Phenotype Ontology. In addition, PhenoGO was evaluated for coding of anatomical and cellular information and assigning the coded phenotypes to the correct GOA; results obtained show that PhenoGO has a precision of 91% and recall of 92%, demonstrating that the PhenoGO NLP system can accurately encode a large number of anatomical and cellular ontologies to GO annotations. The PhenoGO Database may be accessed at www.phenogo.org.    &lt;br /&gt;
* '''Data Input''': Gene Ontology Annotations Files and Medline Abstracts  &lt;br /&gt;
* '''Data Output''': XML file and www.phenogo.org Web Portal&lt;br /&gt;
* '''Implementation Language''': A variety of modules, the web portal is in Java and MySQL, the computational terminology component (phenOS) is written in Perl scripts that queries tables in IBM DB2,  the natural language processing component is written in PROLOG.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 2, Feb 2006  &lt;br /&gt;
* '''Authors''': Yves Lussier and Carol Friedman are the principal investigators. The programmers are Jianrong Li, Lee Sam, and Tara Borlawsky   &lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': Phenotypic integration, computational phenotypes &lt;br /&gt;
* '''URL''': www.phenogo.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MINDY===&lt;br /&gt;
* '''Description''': Given a transcription factor of interest, MINDY uses a large set of gene expression profile data to identify potential post-transcriptional modulators of the transcription factor's activity. MINDY is based on a three-way statistical interaction model that captures the post-transcriptional regulatory event where the ability of a transcription factor to activate/repress its target genes is monotonically controlled by a potential modulator gene. &lt;br /&gt;
* '''Data Input''': Gene expression data in the EXP format, and a user-specified transcription factor of interest&lt;br /&gt;
* '''Data Output''': Lists of the putative modulators and target genes of the transcription factor, and the modulatory interactions involving them&lt;br /&gt;
* '''Implementation Language''': C++ and MATLAB, Java&lt;br /&gt;
* '''Version, Date, Stage''': Stable release, April 2007  &lt;br /&gt;
* '''Authors''': Kai Wang, Ilya Nemenman, Adam Margolin, Riccardo Dalla-Favera, Andrea Califano&lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': gene expression, transcriptional interaction, modulator &lt;br /&gt;
* '''URL''': n/a&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis --&amp;gt; Regulatory/Signaling network reconstruction&lt;br /&gt;
&lt;br /&gt;
===B Cell Interactome===&lt;br /&gt;
* '''Description''': The B cell interactome (BCI)  is a network of protein-protein, protein-DNA and modulatory interactions in human B cells. The network contains known interactions (reported in public databases) and predicted interactions by a Bayesian evidence integration framework which integrates a variety of generic and context specific experimental clues about protein-protein and protein-DNA interactions - such as a large collection of B cell expression profiles - with inferences from different reverse engineering algorithms, such as GeneWays and ARACNE. Modulatory interactions are predicted by MINDY, an algorithm for the prediction of modulators of transcriptional interactions.&lt;br /&gt;
* '''Data Input''': n/a&lt;br /&gt;
* '''Data Output''':  text file of binary interations associated with a probability. &lt;br /&gt;
* '''Implementation Language''': Perl&lt;br /&gt;
* '''Version, Date, Stage''': Version 2, March 2007&lt;br /&gt;
* '''Authors''':  Lefebvre C, Lim WK, Basso K, Dalla Favera R, and Califano A.&lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': Naive Bayes, Mixed-Interaction Network, human B cells.&lt;br /&gt;
* '''URL''': http://amdec-bioinfo.cu-genome.org/html/BCellInteractome.html&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Interaction Modeling&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10206</id>
		<title>SDIWG:NCBC Software Classification MAGNet Examples</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10206"/>
		<updated>2007-05-11T04:22:10Z</updated>

		<summary type="html">&lt;p&gt;Aris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;===DelPhi===&lt;br /&gt;
* '''Description''': DelPhi provides numerical solutions to the Poisson-Boltzmann equation (both linear and nonlinear form) for molecules of arbitrary shape and charge distribution. The current version is fast (the best relaxation parameter is estimated at run time), accurate (calculation of the electrostatic free energy is less dependent on the resolution of the lattice) and can handle extremely high lattice dimensions. It also includes flexible features for assigning different dielectric constants to different regions of space and treating systems containing mixed salt solutions.&lt;br /&gt;
* '''Data Input''': DelPhi takes as input a coordinate file format of a molecule or equivalent data for geometrical objects and/or charge distributions&lt;br /&gt;
* '''Data Output''': electrostatic potential in and around the system&lt;br /&gt;
* '''Implementation Language''': Fortran and C&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''':  E.Alexov,  R.Fine, M.K.Gilson, A.Nicholls, W.Rocchia, K.Sharp, and B. Honig.&lt;br /&gt;
* '''Platforms Tested''': Unix-SGI IRIX, linux, PC (requires Fortran and C compilers), AIX IBM version and Mac.&lt;br /&gt;
* '''License''': Freely available to academia; pay model for commercial users. &lt;br /&gt;
* '''Keywords''': Finite Difference Poisson-Boltzman Solver&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/delphi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Numerical Calculation of Electrostatic Potential&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP===&lt;br /&gt;
* '''Description''': A molecular visualization and analysis program. It is particularly useful for the display and manipulation of the surfaces of molecules and their electrostatic properties.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi&lt;br /&gt;
* '''Data Output''': molecular graphics.&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': v1.3.6 .Stable public release.&lt;br /&gt;
* '''Authors''': Anthony Nicholls and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': SGI machines: irix 5.x and 6.x (INDYs, INDIGOs including Impact, Octane and O2) systems.&lt;br /&gt;
* '''License''':  Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
===Nest===&lt;br /&gt;
* '''Description''': Modeling protein structure based on a sequence-template alignment. The current server works only for modeling with a single template. Part of jackal, which can be downloaded.&lt;br /&gt;
* '''Data Input''': pir and PDB files&lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Xiang, Z. and Honig, B.&lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': modeling, protein structure, sequence-template alignment.&lt;br /&gt;
* '''URL''': http://honiglab.cpmc.columbia.edu/cgi-bin/jackal/nest.cgi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===JACKAL===&lt;br /&gt;
* '''Description''': Jackal is a collection of programs designed for the modeling and analysis of protein structures. Its core program is a versatile homology modeling package nest. JACKAL has the following capabilities: 1) comparative modeling based on single, composite or multiple templates; 2) side-chain prediction; 3) modeling residue mutation, insertion or deletion; 4) loop prediction; 5) structure refinement; 6) reconstruction of protein missing atoms;7) reconstruction of protein missing residues; 8) prediction of hydrogen atoms; 9) fast calculation of solvent accessible surface area; 10) structure superimposition.       &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Version: 1.5 as of Oct, 20, 2002, Stable public release.&lt;br /&gt;
* '''Authors''': Z. Xiang and B. Honig&lt;br /&gt;
* '''Platforms Tested''': SGI 6.5, Intel Linux and Sun solaris&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': Protein Structure Modeling&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/jackal&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===GRASP2===&lt;br /&gt;
* '''Description''': GRASP2 is an updated version of the GRASP program used for macromolecular structure and surface visualization, contains a large number of new features and scientific tools: Enhanced GUI; Structure alignment and domain database scanning; A gaussian surface generator and new surface coloring schemes; Sequence visualization and alignment; Completed work can be stored in &amp;quot;project files;  Among the many objects that can be stored in a project file are views of the structure; defined subsets, surfaces; Direct printing to printers at full printer resolution.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi, sequence alignments.&lt;br /&gt;
* '''Data Output''': molecular graphics, structural alignments.&lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Donald Petrey and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp2&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Molecular Visualization Package&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===PrISM===&lt;br /&gt;
* '''Description''': PrISM is an integrated computational system where computational tools are implemented for protein sequence and structure analysis and modeling.&lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Wang, L, Yang, A. S. &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''':SGI-irix, Intel-linux &lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': protein analysis/modeling &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/PrISM/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Homology Modeling&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Protein-DNA interface alignment===&lt;br /&gt;
* '''Description''': The protein-DNA alignment software allows one to align the interfacial amino acids from two protein-DNA complexes based on the geometric relationship of each amino acid to its local DNA. &lt;br /&gt;
* '''Data Input''': two PDB files that both contain protein-DNA complexes&lt;br /&gt;
* '''Data Output''': The programs will output the aligned residues and their corresponding residue-residue similarity scores, s(i,j).&lt;br /&gt;
* '''Implementation Language''': C++ and Perl&lt;br /&gt;
* '''Version, Date, Stage''':Stable public release. &lt;br /&gt;
* '''Authors''': Siggers, T.W., Silkov, A &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''': Linux&lt;br /&gt;
* '''License''': Freely available to academia &lt;br /&gt;
* '''Keywords''': protein-DNA interface &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/intfc_aln&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Prediction of Side-chain Conformations&lt;br /&gt;
&lt;br /&gt;
===SURFace===&lt;br /&gt;
* '''Description''': SURFace algorithms are programs that calculate solvent accessible surface area and curvature corrected solvent accessible surface area &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': &lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Nicholls, A., Sharp, K., Sridharan, S. and Honig, B.  &lt;br /&gt;
* '''Platforms Tested''': SGI&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': solvent accessible surface area  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/surf/&lt;br /&gt;
* '''Organization''': MAGNet &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Protein Modeling and Classification --&amp;gt; Caculation of Solvent Accessible Area&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Target Explorer===&lt;br /&gt;
* '''Description''': Automated process of prediction of complex regulatory elements for specified set of transcription factors in Drosophila melanogaster genome. Target Explorer is a complex tool with user-friendly self-explanatory Web-interface that allows to user:  1. create customized library of TF binding site matrices based on user defined sets of training sequences; 2. search for new clusters of binding sites for specified set of TFs; 3.extract annotation for potential target genes.    &lt;br /&gt;
* '''Data Input''': genomic sequences&lt;br /&gt;
* '''Data Output''': clusters of known binding sites &lt;br /&gt;
* '''Implementation Language''': perl, cgi&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release. &lt;br /&gt;
* '''Authors''': Sosinsky A, Bonin CP, Mann RS, Honig B.   &lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': prediction of binding sites for transcription factors  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/Target_Explorer/&lt;br /&gt;
* '''Organization''': MAGNet  &lt;br /&gt;
* '''NCBC Ontology Classification''': Atomic --&amp;gt; SoftwareFunction --&amp;gt; Genomic &amp;amp; Phenotypic Analysis--&amp;gt; Sequence Annotation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===MEDUSA and Gorgon===&lt;br /&gt;
* '''Description''': MEDUSA is an algorithm for learning predictive models of transcriptional gene regulation from gene expression and promoter sequence data.  By using a statistical learning approach based on boosing, MEDUSA learns cis regulatory motifs, condition-specific regulators, and regulatory programs that predict the differential expression of target genes.  The regulatory program is specified as an alternating decision tree (ADT).  The Java implementation of MEDUSA will allow a number of visualizations of the regulatory program and other inferred regulatory information, implemented in the accompanying Gorgon tool, including hits of significant and condition-specific motifs along the promoter sequences of target genes and regulatory network figures viewable in Cytoscape.      &lt;br /&gt;
* '''Data Input''': Discretized (up/down/baseline) gene expression data in plain text format, promoter sequences in FASTA format, list of candidate transcriptional regulators and signal transducers in plain text format.&lt;br /&gt;
* '''Data Output''': Regulatory program represented as a Java serialized object file readable by Gorgon and as a human readable XML file.  Gorgon currently generates views of learned PSSMs, positional hits along promoter sequences, and views of the ADT as HTML files, and generates network figures as Cytoscape format files.&lt;br /&gt;
* '''Implementation Language''': Java (prototyped in MATLAB) &lt;br /&gt;
* '''Version, Date, Stage''': Version 2.0, July 2006, pre-release beta version; Version 1.0 (MATLAB), April 2005, stable public release  &lt;br /&gt;
* '''Authors''': David Quigley, Manuel Middendorf, Steve Lianoglou, Anshul Kundaje, Yoav Freund, Chris Wiggins, Christina Leslie   &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux, Mac OS X&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/medusa (MATLAB),http://compbio.sytes.net:8090/medusa (Java beta version)&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===String kernel package===&lt;br /&gt;
* '''Description''':The string kernel package contains implementations for the mismatch and profile string kernels for use with support vector machine (SVM) classifiers for protein sequence classification.  Both kernels compute similarity between protein sequences based on common occurrences of k-length subsequences (&amp;quot;k-mers&amp;quot;) counted with substitutions.  Kernel functions for protein sequence data enable the training of SVMs for a range of prediction problems, in particular protein structural class prediction and remote homology detection.  A version of the Spider MATLAB machine learning package is also bundled with the code, which allows users to train SVMs and evaluate performance on test sets with the packaged software.&lt;br /&gt;
* '''Data Input''': The mismatch kernel requires sequence data in FASTA format. The profile string kernel uses probabilistic profiles, such as those produced by PSI-BLAST, in place of the original sequences.  The Spider SVM implementation requires both the kernel matrix and a label file of binary or multi-class labels for the training data; this data must be loaded into MATLAB variables before using Spider routing.&lt;br /&gt;
* '''Data Output''':The kernel code produces a kernel matrix for the input data in tab-delimited text format.  The Spider package trains SVMs and stores the learns classifier and results from applying the classifier on test data as MATLAB objects.&lt;br /&gt;
* '''Implementation Language''': String kernel code is implemented in C. Spider is a set of object-oriented MATLAB routines. &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.2, September 2004, stable public release &lt;br /&gt;
* '''Authors''': Eleazar Eskin, Rui Kuang, Eugene Ie, Ke Wang, Jason Weston, Bill Noble, Christina Leslie  &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/string-kernels&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MatrixREDUCE===&lt;br /&gt;
* '''Description''': Regulation of gene expression by a transcription factor requires physical interaction between the factor and the DNA, which can be described by astatistical mechanical model. Based on this model, the MatrixREDUCE algorithm uses genome-wide occupancy data for a transcription factor (e.g.ChIP-chip or mRNA expression data) and associated nucleotide sequences to discover the sequence-specific binding affinity of the transcription factor. The sequence specificity of the transcription factor's DNA-binding domain is modeled using a position-specific affinity matrix (PSAM), representing the change in the binding affinity (Kd) whenever a specific position within a reference binding sequence is mutated. The PSAM can be transformed into affinity logo for visualization using the utility program AffinityLogo, and a MatrixREDUCE run can be summarized in an easy-to-navigate webpage using HTMLSummary. &lt;br /&gt;
* '''Data Input''': sequence file in FASTA format; and expression data file in tab-delimited text format.&lt;br /&gt;
* '''Data Output''': PSAMs in numeric and graphical format, parameters of the fitted model, and an HTML summary page.&lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of Numerical Recipes routines.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.0, July 10, 2006, extensively tested in lab.  &lt;br /&gt;
* '''Authors''': Barrett Foat, Xiang-Jun Lu, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': position-specific affinity matrix, binding affinity, cis-regulatory element, expression data, ChIP-chip, transcription factor &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/MatrixREDUCE&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===T-profiler===&lt;br /&gt;
* '''Description''': T-profiler is a web-based tool that uses the t-test to score changes in the average activity of pre-defined groups of genes. The gene groups are defined based on Gene Ontology categorization, ChIP-chip experiments, upstream matches to a consensus transcription factor binding motif, and location on the same chromosome, respectively. If desired, an iterative procedure can be used to select a single, optimal representative from sets of overlapping gene groups. A jack-knife procedure is used to make calculations more robust against outliers. T-profiler makes it possible to interpret microarray data in a way that is both intuitive and statistically rigorous, without the need to combine experiments or choose parameters.   &lt;br /&gt;
* '''Data Input''': Currently, gene expression data from Saccharomyces cerevisiae and Candida albicans are supported. &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': T-profiler is written in PHP, data is managed by a MYSQL database server  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': André Boorsma, Barrett C. Foat, Daniel Vis, Frans Klis, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Web-based application&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': gene expression, transcriptome,  ChIP-chip,  Gene Ontology &lt;br /&gt;
* '''URL''': http://www.t-profiler.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===TranscriptionDetector===&lt;br /&gt;
* '''Description''':A tool for finding probes measuring significantly expressed loci in a genomic array experiment. Given expression data from some tiling array experiment, TranscriptionDetector decides the likelihood that a probe is detecting transcription from the locus in which it resides. Probabilities are assigned by making use of a background signal intensity distribution from a set of negative control probes. This tool is useful for the functional annotation of genomes as it allows for the discovery of novel transcriptional units independently of any genomic annotation.    &lt;br /&gt;
* '''Data Input''': Expression data (GEO or other platforms) and designation of which probes represent negative controls and which are data probes. &lt;br /&gt;
* '''Data Output''': A text file with a list of probes  corresponding to significantly expressed loci. &lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of GSL.  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': Xiang-Jun Lu, Gabor Halasz, Marinus F. van Batenburg  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': tiling arrays, expression, transcriptome &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/TranscriptionDetector/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===PhenoGO===&lt;br /&gt;
* '''Description''': PhenoGO adds phenotypic contextual information to existing associations between gene products and Gene Ontology (GO) terms as specified in GO Annotations (GOA). PhenoGO utilizes an existing Natural Language Processing (NLP) system, called BioMedLEE, an existing knowledge-based phenotype organizer system (PhenOS) in conjunction with MeSH indexing and established biomedical ontologies. The system also encodes the context to identifiers that are associated in different biomedical ontologies, including the UMLS, Cell Ontology, Mouse Anatomy, NCBI taxonomy, GO, and Mammalian Phenotype Ontology. In addition, PhenoGO was evaluated for coding of anatomical and cellular information and assigning the coded phenotypes to the correct GOA; results obtained show that PhenoGO has a precision of 91% and recall of 92%, demonstrating that the PhenoGO NLP system can accurately encode a large number of anatomical and cellular ontologies to GO annotations. The PhenoGO Database may be accessed at www.phenogo.org.    &lt;br /&gt;
* '''Data Input''': Gene Ontology Annotations Files and Medline Abstracts  &lt;br /&gt;
* '''Data Output''': XML file and www.phenogo.org Web Portal&lt;br /&gt;
* '''Implementation Language''': A variety of modules, the web portal is in Java and MySQL, the computational terminology component (phenOS) is written in Perl scripts that queries tables in IBM DB2,  the natural language processing component is written in PROLOG.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 2, Feb 2006  &lt;br /&gt;
* '''Authors''': Yves Lussier and Carol Friedman are the principal investigators. The programmers are Jianrong Li, Lee Sam, and Tara Borlawsky   &lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': Phenotypic integration, computational phenotypes &lt;br /&gt;
* '''URL''': www.phenogo.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10205</id>
		<title>SDIWG:NCBC Software Classification MAGNet Examples</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10205"/>
		<updated>2007-05-11T03:38:21Z</updated>

		<summary type="html">&lt;p&gt;Aris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;===DelPhi===&lt;br /&gt;
* '''Description''': DelPhi provides numerical solutions to the Poisson-Boltzmann equation (both linear and nonlinear form) for molecules of arbitrary shape and charge distribution. The current version is fast (the best relaxation parameter is estimated at run time), accurate (calculation of the electrostatic free energy is less dependent on the resolution of the lattice) and can handle extremely high lattice dimensions. It also includes flexible features for assigning different dielectric constants to different regions of space and treating systems containing mixed salt solutions.&lt;br /&gt;
* '''Data Input''': DelPhi takes as input a coordinate file format of a molecule or equivalent data for geometrical objects and/or charge distributions&lt;br /&gt;
* '''Data Output''': electrostatic potential in and around the system&lt;br /&gt;
* '''Implementation Language''': Fortran and C&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''':  E.Alexov,  R.Fine, M.K.Gilson, A.Nicholls, W.Rocchia, K.Sharp, and B. Honig.&lt;br /&gt;
* '''Platforms Tested''': Unix-SGI IRIX, linux, PC (requires Fortran and C compilers), AIX IBM version and Mac.&lt;br /&gt;
* '''License''': Freely available to academia; pay model for commercial users. &lt;br /&gt;
* '''Keywords''': Finite Difference Poisson-Boltzman Solver&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/delphi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===GRASP===&lt;br /&gt;
* '''Description''': A molecular visualization and analysis program. It is particularly useful for the display and manipulation of the surfaces of molecules and their electrostatic properties.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi&lt;br /&gt;
* '''Data Output''': molecular graphics.&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': v1.3.6 .Stable public release.&lt;br /&gt;
* '''Authors''': Anthony Nicholls and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': SGI machines: irix 5.x and 6.x (INDYs, INDIGOs including Impact, Octane and O2) systems.&lt;br /&gt;
* '''License''':  Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===Nest===&lt;br /&gt;
* '''Description''': Modeling protein structure based on a sequence-template alignment. The current server works only for modeling with a single template. Part of jackal, which can be downloaded.&lt;br /&gt;
* '''Data Input''': pir and PDB files&lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Xiang, Z. and Honig, B.&lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': modeling, protein structure, sequence-template alignment.&lt;br /&gt;
* '''URL''': http://honiglab.cpmc.columbia.edu/cgi-bin/jackal/nest.cgi&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===JACKAL===&lt;br /&gt;
* '''Description''': Jackal is a collection of programs designed for the modeling and analysis of protein structures. Its core program is a versatile homology modeling package nest. JACKAL has the following capabilities: 1) comparative modeling based on single, composite or multiple templates; 2) side-chain prediction; 3) modeling residue mutation, insertion or deletion; 4) loop prediction; 5) structure refinement; 6) reconstruction of protein missing atoms;7) reconstruction of protein missing residues; 8) prediction of hydrogen atoms; 9) fast calculation of solvent accessible surface area; 10) structure superimposition.       &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Version: 1.5 as of Oct, 20, 2002, Stable public release.&lt;br /&gt;
* '''Authors''': Z. Xiang and B. Honig&lt;br /&gt;
* '''Platforms Tested''': SGI 6.5, Intel Linux and Sun solaris&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': Protein Structure Modeling&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/jackal&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===GRASP2===&lt;br /&gt;
* '''Description''': GRASP2 is an updated version of the GRASP program used for macromolecular structure and surface visualization, contains a large number of new features and scientific tools: Enhanced GUI; Structure alignment and domain database scanning; A gaussian surface generator and new surface coloring schemes; Sequence visualization and alignment; Completed work can be stored in &amp;quot;project files;  Among the many objects that can be stored in a project file are views of the structure; defined subsets, surfaces; Direct printing to printers at full printer resolution.&lt;br /&gt;
* '''Data Input''': PDB files, potential maps from DelPhi, sequence alignments.&lt;br /&gt;
* '''Data Output''': molecular graphics, structural alignments.&lt;br /&gt;
* '''Implementation Language''': C++&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Donald Petrey and Barry Honig.&lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': molecular visualization&lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/grasp2&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===PrISM===&lt;br /&gt;
* '''Description''': PrISM is an integrated computational system where computational tools are implemented for protein sequence and structure analysis and modeling.&lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': Fortran&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release&lt;br /&gt;
* '''Authors''': Wang, L, Yang, A. S. &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''':SGI-irix, Intel-linux &lt;br /&gt;
* '''License''': Freely available to academia.&lt;br /&gt;
* '''Keywords''': protein analysis/modeling &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/PrISM/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===Protein-DNA interface alignment===&lt;br /&gt;
* '''Description''': The protein-DNA alignment software allows one to align the interfacial amino acids from two protein-DNA complexes based on the geometric relationship of each amino acid to its local DNA. &lt;br /&gt;
* '''Data Input''': two PDB files that both contain protein-DNA complexes&lt;br /&gt;
* '''Data Output''': The programs will output the aligned residues and their corresponding residue-residue similarity scores, s(i,j).&lt;br /&gt;
* '''Implementation Language''': C++ and Perl&lt;br /&gt;
* '''Version, Date, Stage''':Stable public release. &lt;br /&gt;
* '''Authors''': Siggers, T.W., Silkov, A &amp;amp; Honig, B. &lt;br /&gt;
* '''Platforms Tested''': Linux&lt;br /&gt;
* '''License''': Freely available to academia &lt;br /&gt;
* '''Keywords''': protein-DNA interface &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/programs/intfc_aln&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===SURFace===&lt;br /&gt;
* '''Description''': SURFace algorithms are programs that calculate solvent accessible surface area and curvature corrected solvent accessible surface area &lt;br /&gt;
* '''Data Input''': &lt;br /&gt;
* '''Data Output''': &lt;br /&gt;
* '''Implementation Language''': &lt;br /&gt;
* '''Version, Date, Stage''': Stable public release.&lt;br /&gt;
* '''Authors''': Nicholls, A., Sharp, K., Sridharan, S. and Honig, B.  &lt;br /&gt;
* '''Platforms Tested''': SGI&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': solvent accessible surface area  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/surf/&lt;br /&gt;
* '''Organization''': MAGNet &lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===Target Explorer===&lt;br /&gt;
* '''Description''': Automated process of prediction of complex regulatory elements for specified set of transcription factors in Drosophila melanogaster genome. Target Explorer is a complex tool with user-friendly self-explanatory Web-interface that allows to user:  1. create customized library of TF binding site matrices based on user defined sets of training sequences; 2. search for new clusters of binding sites for specified set of TFs; 3.extract annotation for potential target genes.    &lt;br /&gt;
* '''Data Input''': genomic sequences&lt;br /&gt;
* '''Data Output''': clusters of known binding sites &lt;br /&gt;
* '''Implementation Language''': perl, cgi&lt;br /&gt;
* '''Version, Date, Stage''': Stable public release. &lt;br /&gt;
* '''Authors''': Sosinsky A, Bonin CP, Mann RS, Honig B.   &lt;br /&gt;
* '''Platforms Tested''': platform independent (web based tool)&lt;br /&gt;
* '''License''': Freely available to academia. &lt;br /&gt;
* '''Keywords''': prediction of binding sites for transcription factors  &lt;br /&gt;
* '''URL''': http://trantor.bioc.columbia.edu/Target_Explorer/&lt;br /&gt;
* '''Organization''': MAGNet  &lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MEDUSA and Gorgon===&lt;br /&gt;
* '''Description''': MEDUSA is an algorithm for learning predictive models of transcriptional gene regulation from gene expression and promoter sequence data.  By using a statistical learning approach based on boosing, MEDUSA learns cis regulatory motifs, condition-specific regulators, and regulatory programs that predict the differential expression of target genes.  The regulatory program is specified as an alternating decision tree (ADT).  The Java implementation of MEDUSA will allow a number of visualizations of the regulatory program and other inferred regulatory information, implemented in the accompanying Gorgon tool, including hits of significant and condition-specific motifs along the promoter sequences of target genes and regulatory network figures viewable in Cytoscape.      &lt;br /&gt;
* '''Data Input''': Discretized (up/down/baseline) gene expression data in plain text format, promoter sequences in FASTA format, list of candidate transcriptional regulators and signal transducers in plain text format.&lt;br /&gt;
* '''Data Output''': Regulatory program represented as a Java serialized object file readable by Gorgon and as a human readable XML file.  Gorgon currently generates views of learned PSSMs, positional hits along promoter sequences, and views of the ADT as HTML files, and generates network figures as Cytoscape format files.&lt;br /&gt;
* '''Implementation Language''': Java (prototyped in MATLAB) &lt;br /&gt;
* '''Version, Date, Stage''': Version 2.0, July 2006, pre-release beta version; Version 1.0 (MATLAB), April 2005, stable public release  &lt;br /&gt;
* '''Authors''': David Quigley, Manuel Middendorf, Steve Lianoglou, Anshul Kundaje, Yoav Freund, Chris Wiggins, Christina Leslie   &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux, Mac OS X&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/medusa (MATLAB),http://compbio.sytes.net:8090/medusa (Java beta version)&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===String kernel package===&lt;br /&gt;
* '''Description''':The string kernel package contains implementations for the mismatch and profile string kernels for use with support vector machine (SVM) classifiers for protein sequence classification.  Both kernels compute similarity between protein sequences based on common occurrences of k-length subsequences (&amp;quot;k-mers&amp;quot;) counted with substitutions.  Kernel functions for protein sequence data enable the training of SVMs for a range of prediction problems, in particular protein structural class prediction and remote homology detection.  A version of the Spider MATLAB machine learning package is also bundled with the code, which allows users to train SVMs and evaluate performance on test sets with the packaged software.&lt;br /&gt;
* '''Data Input''': The mismatch kernel requires sequence data in FASTA format. The profile string kernel uses probabilistic profiles, such as those produced by PSI-BLAST, in place of the original sequences.  The Spider SVM implementation requires both the kernel matrix and a label file of binary or multi-class labels for the training data; this data must be loaded into MATLAB variables before using Spider routing.&lt;br /&gt;
* '''Data Output''':The kernel code produces a kernel matrix for the input data in tab-delimited text format.  The Spider package trains SVMs and stores the learns classifier and results from applying the classifier on test data as MATLAB objects.&lt;br /&gt;
* '''Implementation Language''': String kernel code is implemented in C. Spider is a set of object-oriented MATLAB routines. &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.2, September 2004, stable public release &lt;br /&gt;
* '''Authors''': Eleazar Eskin, Rui Kuang, Eugene Ie, Ke Wang, Jason Weston, Bill Noble, Christina Leslie  &lt;br /&gt;
* '''Platforms Tested''': Windows, Linux&lt;br /&gt;
* '''License''': Open source &lt;br /&gt;
* '''Keywords''':  &lt;br /&gt;
* '''URL''': http://www.cs.columbia.edu/compbio/string-kernels&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===MatrixREDUCE===&lt;br /&gt;
* '''Description''': Regulation of gene expression by a transcription factor requires physical interaction between the factor and the DNA, which can be described by astatistical mechanical model. Based on this model, the MatrixREDUCE algorithm uses genome-wide occupancy data for a transcription factor (e.g.ChIP-chip or mRNA expression data) and associated nucleotide sequences to discover the sequence-specific binding affinity of the transcription factor. The sequence specificity of the transcription factor's DNA-binding domain is modeled using a position-specific affinity matrix (PSAM), representing the change in the binding affinity (Kd) whenever a specific position within a reference binding sequence is mutated. The PSAM can be transformed into affinity logo for visualization using the utility program AffinityLogo, and a MatrixREDUCE run can be summarized in an easy-to-navigate webpage using HTMLSummary. &lt;br /&gt;
* '''Data Input''': sequence file in FASTA format; and expression data file in tab-delimited text format.&lt;br /&gt;
* '''Data Output''': PSAMs in numeric and graphical format, parameters of the fitted model, and an HTML summary page.&lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of Numerical Recipes routines.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 1.0, July 10, 2006, extensively tested in lab.  &lt;br /&gt;
* '''Authors''': Barrett Foat, Xiang-Jun Lu, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': position-specific affinity matrix, binding affinity, cis-regulatory element, expression data, ChIP-chip, transcription factor &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/MatrixREDUCE&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===T-profiler===&lt;br /&gt;
* '''Description''': T-profiler is a web-based tool that uses the t-test to score changes in the average activity of pre-defined groups of genes. The gene groups are defined based on Gene Ontology categorization, ChIP-chip experiments, upstream matches to a consensus transcription factor binding motif, and location on the same chromosome, respectively. If desired, an iterative procedure can be used to select a single, optimal representative from sets of overlapping gene groups. A jack-knife procedure is used to make calculations more robust against outliers. T-profiler makes it possible to interpret microarray data in a way that is both intuitive and statistically rigorous, without the need to combine experiments or choose parameters.   &lt;br /&gt;
* '''Data Input''': Currently, gene expression data from Saccharomyces cerevisiae and Candida albicans are supported. &lt;br /&gt;
* '''Data Output''':&lt;br /&gt;
* '''Implementation Language''': T-profiler is written in PHP, data is managed by a MYSQL database server  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': André Boorsma, Barrett C. Foat, Daniel Vis, Frans Klis, Harmen J. Bussemaker  &lt;br /&gt;
* '''Platforms Tested''': Web-based application&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': gene expression, transcriptome,  ChIP-chip,  Gene Ontology &lt;br /&gt;
* '''URL''': http://www.t-profiler.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===TranscriptionDetector===&lt;br /&gt;
* '''Description''':A tool for finding probes measuring significantly expressed loci in a genomic array experiment. Given expression data from some tiling array experiment, TranscriptionDetector decides the likelihood that a probe is detecting transcription from the locus in which it resides. Probabilities are assigned by making use of a background signal intensity distribution from a set of negative control probes. This tool is useful for the functional annotation of genomes as it allows for the discovery of novel transcriptional units independently of any genomic annotation.    &lt;br /&gt;
* '''Data Input''': Expression data (GEO or other platforms) and designation of which probes represent negative controls and which are data probes. &lt;br /&gt;
* '''Data Output''': A text file with a list of probes  corresponding to significantly expressed loci. &lt;br /&gt;
* '''Implementation Language''': ANSI C, making use of GSL.  &lt;br /&gt;
* '''Version, Date, Stage''':   &lt;br /&gt;
* '''Authors''': Xiang-Jun Lu, Gabor Halasz, Marinus F. van Batenburg  &lt;br /&gt;
* '''Platforms Tested''': Linux, Cygwin (Windows), Mac OS X&lt;br /&gt;
* '''License''':  &lt;br /&gt;
* '''Keywords''': tiling arrays, expression, transcriptome &lt;br /&gt;
* '''URL''': http://www.bussemakerlab.org/software/TranscriptionDetector/&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;br /&gt;
&lt;br /&gt;
===PhenoGO===&lt;br /&gt;
* '''Description''': PhenoGO adds phenotypic contextual information to existing associations between gene products and Gene Ontology (GO) terms as specified in GO Annotations (GOA). PhenoGO utilizes an existing Natural Language Processing (NLP) system, called BioMedLEE, an existing knowledge-based phenotype organizer system (PhenOS) in conjunction with MeSH indexing and established biomedical ontologies. The system also encodes the context to identifiers that are associated in different biomedical ontologies, including the UMLS, Cell Ontology, Mouse Anatomy, NCBI taxonomy, GO, and Mammalian Phenotype Ontology. In addition, PhenoGO was evaluated for coding of anatomical and cellular information and assigning the coded phenotypes to the correct GOA; results obtained show that PhenoGO has a precision of 91% and recall of 92%, demonstrating that the PhenoGO NLP system can accurately encode a large number of anatomical and cellular ontologies to GO annotations. The PhenoGO Database may be accessed at www.phenogo.org.    &lt;br /&gt;
* '''Data Input''': Gene Ontology Annotations Files and Medline Abstracts  &lt;br /&gt;
* '''Data Output''': XML file and www.phenogo.org Web Portal&lt;br /&gt;
* '''Implementation Language''': A variety of modules, the web portal is in Java and MySQL, the computational terminology component (phenOS) is written in Perl scripts that queries tables in IBM DB2,  the natural language processing component is written in PROLOG.  &lt;br /&gt;
* '''Version, Date, Stage''': Version 2, Feb 2006  &lt;br /&gt;
* '''Authors''': Yves Lussier and Carol Friedman are the principal investigators. The programmers are Jianrong Li, Lee Sam, and Tara Borlawsky   &lt;br /&gt;
* '''Platforms Tested''': n/a&lt;br /&gt;
* '''License''':  n/a&lt;br /&gt;
* '''Keywords''': Phenotypic integration, computational phenotypes &lt;br /&gt;
* '''URL''': www.phenogo.org&lt;br /&gt;
* '''Organization''': MAGNet&lt;br /&gt;
* '''NCBC Ontology Classification''':&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10165</id>
		<title>SDIWG:NCBC Software Classification MAGNet Examples</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:NCBC_Software_Classification_MAGNet_Examples&amp;diff=10165"/>
		<updated>2007-05-10T14:47:44Z</updated>

		<summary type="html">&lt;p&gt;Aris: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;===DelPhi===&lt;br /&gt;
* Description: DelPhi provides numerical solutions to the Poisson-Boltzmann equation (both linear and nonlinear form) for molecules of arbitrary shape and charge distribution. The current version is fast (the best relaxation parameter is estimated at run time), accurate (calculation of the electrostatic free energy is less dependent on the resolution of the lattice) and can handle extremely high lattice dimensions. It also includes flexible features for assigning different dielectric constants to different regions of space and treating systems containing mixed salt solutions.&lt;br /&gt;
* Input (parameters &amp;amp; Data Types): DelPhi takes as input a coordinate file format of a molecule or equivalent data for geometrical objects and/or charge distributions&lt;br /&gt;
* Output Data (parameters &amp;amp; Data Types): electrostatic potential in and around the system&lt;br /&gt;
* Implementation Language: Fortran and C&lt;br /&gt;
* Version, Date, Stage: Stable public release&lt;br /&gt;
* Authors:  E.Alexov,  R.Fine, M.K.Gilson, A.Nicholls, W.Rocchia, K.Sharp, and B. Honig.&lt;br /&gt;
* Platforms Tested: Unix-SGI IRIX, linux, PC (requires Fortran and C compilers), AIX IBM version and Mac.&lt;br /&gt;
* License: Freely available to academia; pay model for commercial users. &lt;br /&gt;
* Keywords: Finite Difference Poisson-Boltzman Solver&lt;br /&gt;
* URL: http://trantor.bioc.columbia.edu/delphi&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:Meeting_Minutes_20070216&amp;diff=7918</id>
		<title>SDIWG:Meeting Minutes 20070216</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:Meeting_Minutes_20070216&amp;diff=7918"/>
		<updated>2007-02-19T20:48:19Z</updated>

		<summary type="html">&lt;p&gt;Aris: /* Action Items */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''Agenda: NCBC Joint Working Group Meeting'''&lt;br /&gt;
&lt;br /&gt;
[[SDIWG:Software_and_Data_Integration_Working_Group|Top page of SDIWG web site]]&lt;br /&gt;
&lt;br /&gt;
===Notes===&lt;br /&gt;
''Friday, February 16, 2007: 2:30 -- 3:30 PM EST''&amp;lt;br&amp;gt;&lt;br /&gt;
''Please contact Peter Lyster for information lysterp@mail.nih.gov''&amp;lt;br&amp;gt;&lt;br /&gt;
''[http://na-mic.org/Wiki/index.php/SDIWG:Software_Engineering_Body_of_Knowledge_Across_NCBC_Biocomputing_Centers SDIWG Meetings Page]''&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== TCon Agenda ===&lt;br /&gt;
&lt;br /&gt;
* Roll Call/Note-taker (Floratos) (5 min)&lt;br /&gt;
&lt;br /&gt;
* Minutes from most recent SDIWG tcon [[SDIWG:Meeting_Minutes_20061117|Nov. 17, 2006]] (5 min)&lt;br /&gt;
&lt;br /&gt;
* Progress in three working groups&lt;br /&gt;
** '''[[SDIWG:_NCBC_Scientific_Ontologies|Scientific Ontologies]]''' (Zak Kohane, Mark Musen, and Suzi Lewis, 15 min)&lt;br /&gt;
*** TBD&lt;br /&gt;
** '''[[SDIWG:_NCBC_Resource_Yellow_Pages_and_Software_Ontologies|Yellow Pages and Resourceome]]''' (Ivo Dinov, Daniel Rubin, Bill Lorensen, Jonathan Dugan, 15 min)&lt;br /&gt;
*** Past Progress&lt;br /&gt;
**** [http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_XMLtology_2006_OtherArchives Other Attempts to provide archives of SW and tool resources]&lt;br /&gt;
**** Proposed set of '''[http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_XMLtology_Meeting_071806_RequiredFields Required/Minimal]''' Resource Description Fields&lt;br /&gt;
**** [http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_XMLtology_Meeting_071806_ToolSpecificFields '''Optional''' and '''Resource-specific''' description fields]&lt;br /&gt;
**** [http://bioontology.org/projects/ontologies/SoftwareOntology/ Current Ontological Description of the NCBC Tool Yellow-Pages]&lt;br /&gt;
**** [http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_XMLtology_Meeting_103106_ResourceDef Types of CompBio Resources]?&lt;br /&gt;
**** [[SDIWG:NCBC_Software_Classification | Center described 3-5 resources]]?&lt;br /&gt;
*** Status of NCBC '''''[http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_iTools iTools]'''''&lt;br /&gt;
**** V1 (v 0.1, 10/12/06)&lt;br /&gt;
**** V4 (v 0.4, 02/13/07)&lt;br /&gt;
*** Sugested Action Items for the YPR WG:&lt;br /&gt;
**** Each Center should ensure that they have [[SDIWG:NCBC_Software_Classification|6-12 SW Resources described according to these templates]]&lt;br /&gt;
**** Test, comment and revise iTools. How does it compare with other similar efforts?&lt;br /&gt;
** '''[[SDIWG:_NCBC_Systems_Biology|Applications of Systems Biology, Modeling, and Analysis]]''' (Aris Floratos, Brian Athey, Russ Altman, 15 min)&lt;br /&gt;
*** TBD&lt;br /&gt;
&lt;br /&gt;
=== Minutes ===&lt;br /&gt;
&lt;br /&gt;
Note taker: (Aris Floratos)&lt;br /&gt;
&lt;br /&gt;
* Review of 11/17/06 minutes&lt;br /&gt;
** '''[[SDIWG:_NCBC_Resource_Yellow_Pages_and_Software_Ontologies|Yellow Pages and Resourceome]]'''&lt;br /&gt;
** '''[[SDIWG:_NCBC_Scientific_Ontologies|Scientific Ontologies]]'''&lt;br /&gt;
** '''[[SDIWG:_NCBC_Systems_Biology|Applications of Systems Biology, Modeling, and Analysis]]'''&lt;br /&gt;
&lt;br /&gt;
Zak Kohane outlined the process via which the Scientific Ontologies SIG reviewed a select set of ontologies and vocabularies. A summary list along with the a broad assessment of overall &amp;quot;quality&amp;quot; has been compiled by Suzie Lewis and is available at www.berkeleybop.org/sowg/table.cgi. In addition to being an inventory for the vocabularies and ontologies that tbe NCBCs use intenally, the SIG hopes that this list will be a useful starting point to non-experts seeking to identify ontologies/terminologies appropriate for use in their work. Zak recognized that there is a  gap in the set of vocabularies reviewed, namely in the area of imaging and neuroscience and he solicited help in identifying the main players in these domains. Daniel Rubin offered to help in this direction.&lt;br /&gt;
&lt;br /&gt;
In response to questions/suggestions from Gil Ommenn and Brian Athey, it was agreed that the Scientific Ontologies SIG will also produce representative use cases, outlining how ontologies/terminologies are used in practice.&lt;br /&gt;
&lt;br /&gt;
Ivo Dinov reviewed the process that led to the definition of the minimum resource descriptors and Dan Rubin's software classification ontology and proceeded to discuss the iTools framework (www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_iTools) and demonstrate the Java prototype front end. The demo outlined several supported search modes as well as the ability of the tool to &amp;quot;scrape&amp;quot; updates to tool description information from Web pages that follow the format of the Resource descriptions on the SDIWG wiki.&lt;br /&gt;
&lt;br /&gt;
Suzie Lewis observed that the scraping approach useb by the iTools could be used to also process the listing of terminologies produced by the Ontologies SIG and allow then to be presented/searched via the iTools application.&lt;br /&gt;
&lt;br /&gt;
A discussion followed about what is the proper manner to release and start using the iTools framework. Suggestions were made taht NIH should centrally host the database and portal, so as to offer the approrpiate branding and central entry point for the general public to access the NCBC resources. One idea was to develop a NCBC-specific web site (e.g., ncbc.org; or nih.ncbc.org) for that purpose. Peter Lyster said that there have been internal discussion in NIH about this topic and some ideas (such as having a central entry point to the NCBC resource listing from NCBI or some other institute) have been vetted.&lt;br /&gt;
&lt;br /&gt;
Brian Athey and Aris Floratos commented on the upcoming March telecon of the Systems Biology SIG and the idea of developing a suite of software tools comprising resources from mulitple NCBCs which will be leveraged in the context of various DBPs. The question came up of how we achieve a good balance between tool development and tool adoption and how we advertise and promote the various tools, in particular since the adoption of these tools by the community will, to a large extent, determine the impact and usefulness of the NCBC program as whole and its potential for renewed funding. Peter Lyster commented that there is no concerted effort at the time wihtin NIH to promote the NCBC tools, rather there are a number of ongoing initiatives like caBIG, BIRN, etc that are procedding in parallel. One course of action might be to start with having the iTools hosted at CCB and then move gradually towards bringing the repository within NIH as an updated NCBC page on the BISTI site (or elsewhere) is created. Peter also said that he will ask John Whitmarsh to get back to the PIs regarding the current plans for the program evaluation.&lt;br /&gt;
&lt;br /&gt;
=== Attendees ===&lt;br /&gt;
&lt;br /&gt;
* Sherman, Lyster, Dinov (CCB/UCLA), Lorensen, Michael Montegut (NCBO project mgr), Weymouth, Floratos, Jennie Larkin (NHLBI), Zak &amp;amp; Susanne, Jenkins (NLM), Vivien Bonai, Athey, Kennedy, Mike Mendis, Karen Skinner/NIDA, Musen, Chris Mungall (NCBO), German Cavelier (NIMH), Suzanna Lewis&lt;br /&gt;
* Note-taker: [Suggested order of note takers for future meetings: Floratos (this one); Rubin; Jags; Dinov; Chueh; Sherman; Schroeder]&lt;br /&gt;
&lt;br /&gt;
=== Action Items ===&lt;br /&gt;
* Dan Rubin to point our for the Scientific Ontologies SIG the key imaging/neuroscience ontologies.&lt;br /&gt;
* Ontologies SIG to produce use cases describing practical uses of ontologies/vocabularies.&lt;br /&gt;
* Peter Lyster (or/and others?) to bring up as an agenda item to the NCBC project team meeting the issue of hosting the Yellow Pages database centrally within NIH.&lt;br /&gt;
* Peter Lyster to ask John Whitmarsh to communicate to the PIs the status of plans for the NCBC evaluation program.&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:Meeting_Minutes_20070216&amp;diff=7917</id>
		<title>SDIWG:Meeting Minutes 20070216</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:Meeting_Minutes_20070216&amp;diff=7917"/>
		<updated>2007-02-19T20:47:47Z</updated>

		<summary type="html">&lt;p&gt;Aris: /* Action Items */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''Agenda: NCBC Joint Working Group Meeting'''&lt;br /&gt;
&lt;br /&gt;
[[SDIWG:Software_and_Data_Integration_Working_Group|Top page of SDIWG web site]]&lt;br /&gt;
&lt;br /&gt;
===Notes===&lt;br /&gt;
''Friday, February 16, 2007: 2:30 -- 3:30 PM EST''&amp;lt;br&amp;gt;&lt;br /&gt;
''Please contact Peter Lyster for information lysterp@mail.nih.gov''&amp;lt;br&amp;gt;&lt;br /&gt;
''[http://na-mic.org/Wiki/index.php/SDIWG:Software_Engineering_Body_of_Knowledge_Across_NCBC_Biocomputing_Centers SDIWG Meetings Page]''&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== TCon Agenda ===&lt;br /&gt;
&lt;br /&gt;
* Roll Call/Note-taker (Floratos) (5 min)&lt;br /&gt;
&lt;br /&gt;
* Minutes from most recent SDIWG tcon [[SDIWG:Meeting_Minutes_20061117|Nov. 17, 2006]] (5 min)&lt;br /&gt;
&lt;br /&gt;
* Progress in three working groups&lt;br /&gt;
** '''[[SDIWG:_NCBC_Scientific_Ontologies|Scientific Ontologies]]''' (Zak Kohane, Mark Musen, and Suzi Lewis, 15 min)&lt;br /&gt;
*** TBD&lt;br /&gt;
** '''[[SDIWG:_NCBC_Resource_Yellow_Pages_and_Software_Ontologies|Yellow Pages and Resourceome]]''' (Ivo Dinov, Daniel Rubin, Bill Lorensen, Jonathan Dugan, 15 min)&lt;br /&gt;
*** Past Progress&lt;br /&gt;
**** [http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_XMLtology_2006_OtherArchives Other Attempts to provide archives of SW and tool resources]&lt;br /&gt;
**** Proposed set of '''[http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_XMLtology_Meeting_071806_RequiredFields Required/Minimal]''' Resource Description Fields&lt;br /&gt;
**** [http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_XMLtology_Meeting_071806_ToolSpecificFields '''Optional''' and '''Resource-specific''' description fields]&lt;br /&gt;
**** [http://bioontology.org/projects/ontologies/SoftwareOntology/ Current Ontological Description of the NCBC Tool Yellow-Pages]&lt;br /&gt;
**** [http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_XMLtology_Meeting_103106_ResourceDef Types of CompBio Resources]?&lt;br /&gt;
**** [[SDIWG:NCBC_Software_Classification | Center described 3-5 resources]]?&lt;br /&gt;
*** Status of NCBC '''''[http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_iTools iTools]'''''&lt;br /&gt;
**** V1 (v 0.1, 10/12/06)&lt;br /&gt;
**** V4 (v 0.4, 02/13/07)&lt;br /&gt;
*** Sugested Action Items for the YPR WG:&lt;br /&gt;
**** Each Center should ensure that they have [[SDIWG:NCBC_Software_Classification|6-12 SW Resources described according to these templates]]&lt;br /&gt;
**** Test, comment and revise iTools. How does it compare with other similar efforts?&lt;br /&gt;
** '''[[SDIWG:_NCBC_Systems_Biology|Applications of Systems Biology, Modeling, and Analysis]]''' (Aris Floratos, Brian Athey, Russ Altman, 15 min)&lt;br /&gt;
*** TBD&lt;br /&gt;
&lt;br /&gt;
=== Minutes ===&lt;br /&gt;
&lt;br /&gt;
Note taker: (Aris Floratos)&lt;br /&gt;
&lt;br /&gt;
* Review of 11/17/06 minutes&lt;br /&gt;
** '''[[SDIWG:_NCBC_Resource_Yellow_Pages_and_Software_Ontologies|Yellow Pages and Resourceome]]'''&lt;br /&gt;
** '''[[SDIWG:_NCBC_Scientific_Ontologies|Scientific Ontologies]]'''&lt;br /&gt;
** '''[[SDIWG:_NCBC_Systems_Biology|Applications of Systems Biology, Modeling, and Analysis]]'''&lt;br /&gt;
&lt;br /&gt;
Zak Kohane outlined the process via which the Scientific Ontologies SIG reviewed a select set of ontologies and vocabularies. A summary list along with the a broad assessment of overall &amp;quot;quality&amp;quot; has been compiled by Suzie Lewis and is available at www.berkeleybop.org/sowg/table.cgi. In addition to being an inventory for the vocabularies and ontologies that tbe NCBCs use intenally, the SIG hopes that this list will be a useful starting point to non-experts seeking to identify ontologies/terminologies appropriate for use in their work. Zak recognized that there is a  gap in the set of vocabularies reviewed, namely in the area of imaging and neuroscience and he solicited help in identifying the main players in these domains. Daniel Rubin offered to help in this direction.&lt;br /&gt;
&lt;br /&gt;
In response to questions/suggestions from Gil Ommenn and Brian Athey, it was agreed that the Scientific Ontologies SIG will also produce representative use cases, outlining how ontologies/terminologies are used in practice.&lt;br /&gt;
&lt;br /&gt;
Ivo Dinov reviewed the process that led to the definition of the minimum resource descriptors and Dan Rubin's software classification ontology and proceeded to discuss the iTools framework (www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_iTools) and demonstrate the Java prototype front end. The demo outlined several supported search modes as well as the ability of the tool to &amp;quot;scrape&amp;quot; updates to tool description information from Web pages that follow the format of the Resource descriptions on the SDIWG wiki.&lt;br /&gt;
&lt;br /&gt;
Suzie Lewis observed that the scraping approach useb by the iTools could be used to also process the listing of terminologies produced by the Ontologies SIG and allow then to be presented/searched via the iTools application.&lt;br /&gt;
&lt;br /&gt;
A discussion followed about what is the proper manner to release and start using the iTools framework. Suggestions were made taht NIH should centrally host the database and portal, so as to offer the approrpiate branding and central entry point for the general public to access the NCBC resources. One idea was to develop a NCBC-specific web site (e.g., ncbc.org; or nih.ncbc.org) for that purpose. Peter Lyster said that there have been internal discussion in NIH about this topic and some ideas (such as having a central entry point to the NCBC resource listing from NCBI or some other institute) have been vetted.&lt;br /&gt;
&lt;br /&gt;
Brian Athey and Aris Floratos commented on the upcoming March telecon of the Systems Biology SIG and the idea of developing a suite of software tools comprising resources from mulitple NCBCs which will be leveraged in the context of various DBPs. The question came up of how we achieve a good balance between tool development and tool adoption and how we advertise and promote the various tools, in particular since the adoption of these tools by the community will, to a large extent, determine the impact and usefulness of the NCBC program as whole and its potential for renewed funding. Peter Lyster commented that there is no concerted effort at the time wihtin NIH to promote the NCBC tools, rather there are a number of ongoing initiatives like caBIG, BIRN, etc that are procedding in parallel. One course of action might be to start with having the iTools hosted at CCB and then move gradually towards bringing the repository within NIH as an updated NCBC page on the BISTI site (or elsewhere) is created. Peter also said that he will ask John Whitmarsh to get back to the PIs regarding the current plans for the program evaluation.&lt;br /&gt;
&lt;br /&gt;
=== Attendees ===&lt;br /&gt;
&lt;br /&gt;
* Sherman, Lyster, Dinov (CCB/UCLA), Lorensen, Michael Montegut (NCBO project mgr), Weymouth, Floratos, Jennie Larkin (NHLBI), Zak &amp;amp; Susanne, Jenkins (NLM), Vivien Bonai, Athey, Kennedy, Mike Mendis, Karen Skinner/NIDA, Musen, Chris Mungall (NCBO), German Cavelier (NIMH), Suzanna Lewis&lt;br /&gt;
* Note-taker: [Suggested order of note takers for future meetings: Floratos (this one); Rubin; Jags; Dinov; Chueh; Sherman; Schroeder]&lt;br /&gt;
&lt;br /&gt;
=== Action Items ===&lt;br /&gt;
* Dan Rubin to point our for the Ontology SIG the key imaging/neuroscience ontologies.&lt;br /&gt;
* Ontology SIG to produce use cases describing practical uses of ontologies/vocabularies.&lt;br /&gt;
* Peter Lyster (or/and others?) to bring up as an agenda item to the NCBC project team meeting the issue of hosting the Yellow Pages database centrally within NIH.&lt;br /&gt;
* Peter Lyster to ask John Whitmarsh to communicate to the PIs the status of plans for the NCBC evaluation program.&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
	<entry>
		<id>https://www.na-mic.org/w/index.php?title=SDIWG:Meeting_Minutes_20070216&amp;diff=7916</id>
		<title>SDIWG:Meeting Minutes 20070216</title>
		<link rel="alternate" type="text/html" href="https://www.na-mic.org/w/index.php?title=SDIWG:Meeting_Minutes_20070216&amp;diff=7916"/>
		<updated>2007-02-19T20:47:04Z</updated>

		<summary type="html">&lt;p&gt;Aris: /* Minutes */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''Agenda: NCBC Joint Working Group Meeting'''&lt;br /&gt;
&lt;br /&gt;
[[SDIWG:Software_and_Data_Integration_Working_Group|Top page of SDIWG web site]]&lt;br /&gt;
&lt;br /&gt;
===Notes===&lt;br /&gt;
''Friday, February 16, 2007: 2:30 -- 3:30 PM EST''&amp;lt;br&amp;gt;&lt;br /&gt;
''Please contact Peter Lyster for information lysterp@mail.nih.gov''&amp;lt;br&amp;gt;&lt;br /&gt;
''[http://na-mic.org/Wiki/index.php/SDIWG:Software_Engineering_Body_of_Knowledge_Across_NCBC_Biocomputing_Centers SDIWG Meetings Page]''&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== TCon Agenda ===&lt;br /&gt;
&lt;br /&gt;
* Roll Call/Note-taker (Floratos) (5 min)&lt;br /&gt;
&lt;br /&gt;
* Minutes from most recent SDIWG tcon [[SDIWG:Meeting_Minutes_20061117|Nov. 17, 2006]] (5 min)&lt;br /&gt;
&lt;br /&gt;
* Progress in three working groups&lt;br /&gt;
** '''[[SDIWG:_NCBC_Scientific_Ontologies|Scientific Ontologies]]''' (Zak Kohane, Mark Musen, and Suzi Lewis, 15 min)&lt;br /&gt;
*** TBD&lt;br /&gt;
** '''[[SDIWG:_NCBC_Resource_Yellow_Pages_and_Software_Ontologies|Yellow Pages and Resourceome]]''' (Ivo Dinov, Daniel Rubin, Bill Lorensen, Jonathan Dugan, 15 min)&lt;br /&gt;
*** Past Progress&lt;br /&gt;
**** [http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_XMLtology_2006_OtherArchives Other Attempts to provide archives of SW and tool resources]&lt;br /&gt;
**** Proposed set of '''[http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_XMLtology_Meeting_071806_RequiredFields Required/Minimal]''' Resource Description Fields&lt;br /&gt;
**** [http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_XMLtology_Meeting_071806_ToolSpecificFields '''Optional''' and '''Resource-specific''' description fields]&lt;br /&gt;
**** [http://bioontology.org/projects/ontologies/SoftwareOntology/ Current Ontological Description of the NCBC Tool Yellow-Pages]&lt;br /&gt;
**** [http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_XMLtology_Meeting_103106_ResourceDef Types of CompBio Resources]?&lt;br /&gt;
**** [[SDIWG:NCBC_Software_Classification | Center described 3-5 resources]]?&lt;br /&gt;
*** Status of NCBC '''''[http://www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_iTools iTools]'''''&lt;br /&gt;
**** V1 (v 0.1, 10/12/06)&lt;br /&gt;
**** V4 (v 0.4, 02/13/07)&lt;br /&gt;
*** Sugested Action Items for the YPR WG:&lt;br /&gt;
**** Each Center should ensure that they have [[SDIWG:NCBC_Software_Classification|6-12 SW Resources described according to these templates]]&lt;br /&gt;
**** Test, comment and revise iTools. How does it compare with other similar efforts?&lt;br /&gt;
** '''[[SDIWG:_NCBC_Systems_Biology|Applications of Systems Biology, Modeling, and Analysis]]''' (Aris Floratos, Brian Athey, Russ Altman, 15 min)&lt;br /&gt;
*** TBD&lt;br /&gt;
&lt;br /&gt;
=== Minutes ===&lt;br /&gt;
&lt;br /&gt;
Note taker: (Aris Floratos)&lt;br /&gt;
&lt;br /&gt;
* Review of 11/17/06 minutes&lt;br /&gt;
** '''[[SDIWG:_NCBC_Resource_Yellow_Pages_and_Software_Ontologies|Yellow Pages and Resourceome]]'''&lt;br /&gt;
** '''[[SDIWG:_NCBC_Scientific_Ontologies|Scientific Ontologies]]'''&lt;br /&gt;
** '''[[SDIWG:_NCBC_Systems_Biology|Applications of Systems Biology, Modeling, and Analysis]]'''&lt;br /&gt;
&lt;br /&gt;
Zak Kohane outlined the process via which the Scientific Ontologies SIG reviewed a select set of ontologies and vocabularies. A summary list along with the a broad assessment of overall &amp;quot;quality&amp;quot; has been compiled by Suzie Lewis and is available at www.berkeleybop.org/sowg/table.cgi. In addition to being an inventory for the vocabularies and ontologies that tbe NCBCs use intenally, the SIG hopes that this list will be a useful starting point to non-experts seeking to identify ontologies/terminologies appropriate for use in their work. Zak recognized that there is a  gap in the set of vocabularies reviewed, namely in the area of imaging and neuroscience and he solicited help in identifying the main players in these domains. Daniel Rubin offered to help in this direction.&lt;br /&gt;
&lt;br /&gt;
In response to questions/suggestions from Gil Ommenn and Brian Athey, it was agreed that the Scientific Ontologies SIG will also produce representative use cases, outlining how ontologies/terminologies are used in practice.&lt;br /&gt;
&lt;br /&gt;
Ivo Dinov reviewed the process that led to the definition of the minimum resource descriptors and Dan Rubin's software classification ontology and proceeded to discuss the iTools framework (www.loni.ucla.edu/twiki/bin/view/CCB/NCBC_ToolIntegration_iTools) and demonstrate the Java prototype front end. The demo outlined several supported search modes as well as the ability of the tool to &amp;quot;scrape&amp;quot; updates to tool description information from Web pages that follow the format of the Resource descriptions on the SDIWG wiki.&lt;br /&gt;
&lt;br /&gt;
Suzie Lewis observed that the scraping approach useb by the iTools could be used to also process the listing of terminologies produced by the Ontologies SIG and allow then to be presented/searched via the iTools application.&lt;br /&gt;
&lt;br /&gt;
A discussion followed about what is the proper manner to release and start using the iTools framework. Suggestions were made taht NIH should centrally host the database and portal, so as to offer the approrpiate branding and central entry point for the general public to access the NCBC resources. One idea was to develop a NCBC-specific web site (e.g., ncbc.org; or nih.ncbc.org) for that purpose. Peter Lyster said that there have been internal discussion in NIH about this topic and some ideas (such as having a central entry point to the NCBC resource listing from NCBI or some other institute) have been vetted.&lt;br /&gt;
&lt;br /&gt;
Brian Athey and Aris Floratos commented on the upcoming March telecon of the Systems Biology SIG and the idea of developing a suite of software tools comprising resources from mulitple NCBCs which will be leveraged in the context of various DBPs. The question came up of how we achieve a good balance between tool development and tool adoption and how we advertise and promote the various tools, in particular since the adoption of these tools by the community will, to a large extent, determine the impact and usefulness of the NCBC program as whole and its potential for renewed funding. Peter Lyster commented that there is no concerted effort at the time wihtin NIH to promote the NCBC tools, rather there are a number of ongoing initiatives like caBIG, BIRN, etc that are procedding in parallel. One course of action might be to start with having the iTools hosted at CCB and then move gradually towards bringing the repository within NIH as an updated NCBC page on the BISTI site (or elsewhere) is created. Peter also said that he will ask John Whitmarsh to get back to the PIs regarding the current plans for the program evaluation.&lt;br /&gt;
&lt;br /&gt;
=== Attendees ===&lt;br /&gt;
&lt;br /&gt;
* Sherman, Lyster, Dinov (CCB/UCLA), Lorensen, Michael Montegut (NCBO project mgr), Weymouth, Floratos, Jennie Larkin (NHLBI), Zak &amp;amp; Susanne, Jenkins (NLM), Vivien Bonai, Athey, Kennedy, Mike Mendis, Karen Skinner/NIDA, Musen, Chris Mungall (NCBO), German Cavelier (NIMH), Suzanna Lewis&lt;br /&gt;
* Note-taker: [Suggested order of note takers for future meetings: Floratos (this one); Rubin; Jags; Dinov; Chueh; Sherman; Schroeder]&lt;br /&gt;
&lt;br /&gt;
=== Action Items ===&lt;/div&gt;</summary>
		<author><name>Aris</name></author>
		
	</entry>
</feed>