Pathway and Gene to Pubmed
This workflow takes in a list of gene names and searches the PubMed database for corresponding articles. Any matches to the genes are then retrieved (abstracts only). These abstracts are then returned to the user.
Preview
Run
Run this Workflow in the Taverna Workbench...
Option 1:
Copy and paste this link into File > 'Open workflow location...'
http://www.myexperiment.org/workflows/1846/download?version=1
[ More Info ]
Taverna is available from http://taverna.sourceforge.net/
If you are having problems downloading it in Taverna, you may need to provide your username and password in the URL so that Taverna can access the Workflow:
Replace http:// in the link above with http://yourusername:yourpassword@
Workflow Components
Paul Fisher |
Gene to Pubmed |
Pathway and Gene to Pubmed |
This workflow takes in a list of gene names and searches the PubMed database for corresponding articles. Any matches to the genes are then retrieved (abstracts only). These abstracts are then returned to the user. |
None
Name | Description |
---|---|
gene_names | |
Phenotype_search_term | |
pathway_descriptions |
Name | Type | Description |
---|---|---|
regex | stringconstant |
Value\n |
remove_nulls | beanshell |
ScriptString[] split = input.split("\n"); Vector nonEmpty = new Vector(); for (int i = 0; i < split.length; i++){ if (!(split[i].equals(""))) { nonEmpty.add(split[i].trim()); } } String[] non_empty = new String[nonEmpty.size()]; for (int i = 0; i < non_empty.length; i ++) { non_empty[i] = nonEmpty.elementAt(i); } String output = ""; for (int i = 0; i < non_empty.length; i++) { output = output + (String) (non_empty[i] + "\n"); } |
gene_and_abstract | beanshell |
ScriptString[] split = abstracts.split("\n"); String pathway_name = pathway; Vector nonEmpty = new Vector(); for (int i = 0; i < split.length; i++) { String trimmed = split[i].trim(); nonEmpty.add(trimmed); } String output = ">> " + pathway_name + "\n"; for (int i = 0; i < nonEmpty.size(); i++) { output = output + (String) (nonEmpty.elementAt(i) + "\n"); } |
split_search_terms | localworker |
ScriptList split = new ArrayList(); if (!string.equals("")) { String regexString = ","; if (regex != void) { regexString = regex; } String[] result = string.split(regexString); for (int i = 0; i < result.length; i++) { split.add(result[i]); } } |
merge_outputs_2 | localworker |
ScriptString seperatorString = "\n"; if (seperator != void) { seperatorString = seperator; } StringBuffer sb = new StringBuffer(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); sb.append(item); if (i.hasNext()) { sb.append(seperatorString); } } concatenated = sb.toString(); |
Search_PubMed | workflow | |
clean_text | workflow | |
regular_expression | stringconstant |
Value\n |
Remove_duplicate_genes | localworker |
ScriptList strippedlist = new ArrayList(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); if (strippedlist.contains(item) == false) { strippedlist.add(item); } } |
split_gene_names | localworker |
ScriptList split = new ArrayList(); if (!string.equals("")) { String regexString = ","; if (regex != void) { regexString = regex; } String[] result = string.split(regexString); for (int i = 0; i < result.length; i++) { split.add(result[i]); } } |
remove_Nulls_2 | beanshell |
ScriptString[] split = input.split("\n"); Vector nonEmpty = new Vector(); for (int i = 0; i < split.length; i++){ if (!(split[i].equals(""))) { nonEmpty.add(split[i].trim()); } } String[] non_empty = new String[nonEmpty.size()]; for (int i = 0; i < non_empty.length; i ++) { non_empty[i] = nonEmpty.elementAt(i); } String output = ""; for (int i = 0; i < non_empty.length; i++) { output = output + (String) (non_empty[i] + "\n"); } |
Merge_string_list_to_string | localworker |
ScriptString seperatorString = "\n"; if (seperator != void) { seperatorString = seperator; } StringBuffer sb = new StringBuffer(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); sb.append(item); if (i.hasNext()) { sb.append(seperatorString); } } concatenated = sb.toString(); |
xpath | stringconstant |
Value/*[local-name(.)='eSearchResult']/*[local-name(.)='IdList']/*[local-name(.)='Id'] |
run_eSearch | wsdl |
Wsdlhttp://eutils.ncbi.nlm.nih.gov/entrez/eutils/soap/eutils.wsdlWsdl Operationrun_eSearch |
min_date | stringconstant |
Value2000/01/01 |
extractPMID | localworker |
Scriptimport org.dom4j.Document; import org.dom4j.Node; import org.dom4j.io.SAXReader; SAXReader reader = new SAXReader(false); reader.setIncludeInternalDTDDeclarations(false); reader.setIncludeExternalDTDDeclarations(false); Document document = reader.read(new StringReader(xmltext)); List nodelist = document.selectNodes(xpath); // Process the elements in the nodelist ArrayList outputList = new ArrayList(); ArrayList outputXmlList = new ArrayList(); String val = null; String xmlVal = null; for (Iterator iter = nodelist.iterator(); iter.hasNext();) { Node element = (Node) iter.next(); xmlVal = element.asXML(); val = element.getStringValue(); if (val != null && !val.equals("")) { outputList.add(val); outputXmlList.add(xmlVal); } } List nodelist=outputList; List nodelistAsXML=outputXmlList; |
remove_Nulls_4 | beanshell |
ScriptString[] split = input.split("\n"); Vector nonEmpty = new Vector(); for (int i = 0; i < split.length; i++){ if (!(split[i].equals(""))) { nonEmpty.add(split[i].trim()); } } String[] non_empty = new String[nonEmpty.size()]; for (int i = 0; i < non_empty.length; i ++) { non_empty[i] = nonEmpty.elementAt(i); } String output = ""; for (int i = 0; i < non_empty.length; i++) { output = output + (String) (non_empty[i] + "\n"); } |
concat_abstract_ids | beanshell |
ScriptString id = id.trim(); String abstract_text = abstract_text.trim(); String[] abstract_array = abstract_text.split("\n"); String abstract_amended = ""; if(abstract_array.length > 1) { abstract_amended = abstract_array[0] + " "; for(int i = 1; i < abstract_array.length; i++) { abstract_amended = abstract_amended + abstract_array[i] + " "; } } else { abstract_amended = abstract_array[0]; } String date_text = date_text.trim(); String output = ""; output = id + "\t" + date_text + "\t" + abstract_amended; |
max_return_phenotype | stringconstant |
Value5000 |
pubmed_database | stringconstant |
Valuepubmed |
merge_phenotype_abstracts | beanshell |
ScriptString[] split = abstracts.split("\n"); String phenotype_term = phenotype.trim(); Vector nonEmpty = new Vector(); for (int i = 0; i < split.length; i++) { String trimmed = split[i].trim(); // String mytext = split[i].substring(split[i].indexOf(0), split[i].indexOf(" AND ")); nonEmpty.add(trimmed); } String output = ">> " + phenotype_term + "\n"; for (int i = 0; i < nonEmpty.size(); i++) { output = output + (String) (nonEmpty.elementAt(i) + "\n"); } |
merge_abstracts | localworker |
ScriptString seperatorString = "\n"; if (seperator != void) { seperatorString = seperator; } StringBuffer sb = new StringBuffer(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); sb.append(item); if (i.hasNext()) { sb.append(seperatorString); } } concatenated = sb.toString(); |
max_date | stringconstant |
Value2011/01/31 |
parametersXML_eSearch | xmlsplitter | |
merge_abstract_ids | localworker |
ScriptString seperatorString = "\n"; if (seperator != void) { seperatorString = seperator; } StringBuffer sb = new StringBuffer(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); sb.append(item); if (i.hasNext()) { sb.append(seperatorString); } } concatenated = sb.toString(); |
merge_dates | localworker |
ScriptString seperatorString = "\n"; if (seperator != void) { seperatorString = seperator; } StringBuffer sb = new StringBuffer(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); sb.append(item); if (i.hasNext()) { sb.append(seperatorString); } } concatenated = sb.toString(); |
Fetch_Abstracts | workflow | |
clean_text_3 | workflow | |
Encode_byte_to_base64 | localworker |
Scriptimport org.apache.commons.codec.binary.Base64; base64 = new String(Base64.encodeBase64(bytes)); |
merge_strings_2 | localworker |
ScriptString seperatorString = "\n"; if (seperator != void) { seperatorString = seperator; } StringBuffer sb = new StringBuffer(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); sb.append(item); if (i.hasNext()) { sb.append(seperatorString); } } concatenated = sb.toString(); |
regex_3 | stringconstant |
Value\n |
xpath_2 | stringconstant |
Value/*[local-name(.)='generateTerminologyResponse']/*[local-name(.)='return'] |
extract_abstracts | beanshell |
ScriptString[] split = input.split("\n"); Vector nonEmpty = new Vector(); for (int i = 1; i < split.length; i++) { String trimmed = split[i].trim(); String[] split_2 = trimmed.split("\t"); if(split_2.length == 3) { nonEmpty.add(split_2[2]); } } String output_search = split[0] + "\n"; String output = ""; for (int i = 0; i < nonEmpty.size(); i++) { output = output + (String) (nonEmpty.elementAt(i) + "\n"); } |
merge_search_and_terms | beanshell |
ScriptString term_input = terms.trim(); String search_input = search_term.trim(); String output = ""; output = search_input + "\n" + term_input; |
extract_Terms_2 | localworker |
Scriptimport org.dom4j.Document; import org.dom4j.Node; import org.dom4j.io.SAXReader; SAXReader reader = new SAXReader(false); reader.setIncludeInternalDTDDeclarations(false); reader.setIncludeExternalDTDDeclarations(false); Document document = reader.read(new StringReader(xmltext)); List nodelist = document.selectNodes(xpath); // Process the elements in the nodelist ArrayList outputList = new ArrayList(); ArrayList outputXmlList = new ArrayList(); String val = null; String xmlVal = null; for (Iterator iter = nodelist.iterator(); iter.hasNext();) { Node element = (Node) iter.next(); xmlVal = element.asXML(); val = element.getStringValue(); if (val != null && !val.equals("")) { outputList.add(val); outputXmlList.add(xmlVal); } } List nodelist=outputList; List nodelistAsXML=outputXmlList; |
merge_strings | localworker |
ScriptString seperatorString = "\n"; if (seperator != void) { seperatorString = seperator; } StringBuffer sb = new StringBuffer(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); sb.append(item); if (i.hasNext()) { sb.append(seperatorString); } } concatenated = sb.toString(); |
remove_Nulls_5 | beanshell |
ScriptString[] split = input.split("\n"); Vector nonEmpty = new Vector(); for (int i = 0; i < split.length; i++){ if (!(split[i].equals(""))) { nonEmpty.add(split[i].trim()); } } String[] non_empty = new String[nonEmpty.size()]; for (int i = 0; i < non_empty.length; i ++) { non_empty[i] = nonEmpty.elementAt(i); } String output = ""; for (int i = 0; i < non_empty.length; i++) { output = output + (String) (non_empty[i] + "\n"); } |
Remove_duplicate_strings_2 | localworker |
ScriptList strippedlist = new ArrayList(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); if (strippedlist.contains(item) == false) { strippedlist.add(item); } } |
split_by_regex | localworker |
ScriptList split = new ArrayList(); if (!string.equals("")) { String regexString = ","; if (regex != void) { regexString = regex; } String[] result = string.split(regexString); for (int i = 0; i < result.length; i++) { split.add(result[i]); } } |
clean_text_4 | soaplab |
Endpointhttp://phoebus.cs.man.ac.uk:1977/axis/services/text_mining.clean_text |
pubmed_database_2 | stringconstant |
Valuepubmed |
xpath_3 | stringconstant |
Value/*[local-name(.)='eSearchResult']/*[local-name(.)='Count'] |
count | stringconstant |
Valuecount |
xpath_count | stringconstant |
Value/*[local-name(.)='eInfoResult']/*[local-name(.)='DbInfo']/*[local-name(.)='Count'] |
eSearch_database | stringconstant |
Valuepubmed |
regular_expression_2 | stringconstant |
Value\n |
extract_terms_3 | beanshell |
ScriptString[] split = input.split("\n"); Vector nonEmpty = new Vector(); for (int i = 1; i < split.length; i++) { String trimmed = split[i].trim(); // if((trimmed.contains("=")) || (trimmed.contains("-"))) // { // next; // } // else // { // String[] trimmed_array = trimmed.split("\t"); // String term = trimmed_array[0]; nonEmpty.add(trimmed); // } } String output = ""; for (int i = 0; i < nonEmpty.size(); i++) { output = output + (String) (nonEmpty.elementAt(i) + "\n"); } |
merge_term_count | beanshell |
ScriptString term_input = term.trim(); String count_input = count.trim(); String output = ""; output = term_input + "\t" + count_input; |
split_extracted_terms | localworker |
ScriptList split = new ArrayList(); if (!string.equals("")) { String regexString = ","; if (regex != void) { regexString = regex; } String[] result = string.split(regexString); for (int i = 0; i < result.length; i++) { split.add(result[i]); } } |
merge_pubmed_count | localworker |
ScriptString seperatorString = "\n"; if (seperator != void) { seperatorString = seperator; } StringBuffer sb = new StringBuffer(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); sb.append(item); if (i.hasNext()) { sb.append(seperatorString); } } concatenated = sb.toString(); |
merge_extracted | localworker |
ScriptString seperatorString = "\n"; if (seperator != void) { seperatorString = seperator; } StringBuffer sb = new StringBuffer(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); sb.append(item); if (i.hasNext()) { sb.append(seperatorString); } } concatenated = sb.toString(); |
merge_list | localworker |
ScriptString seperatorString = "\n"; if (seperator != void) { seperatorString = seperator; } StringBuffer sb = new StringBuffer(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); sb.append(item); if (i.hasNext()) { sb.append(seperatorString); } } concatenated = sb.toString(); |
extractCount_2 | localworker |
Scriptimport org.dom4j.Document; import org.dom4j.Node; import org.dom4j.io.SAXReader; SAXReader reader = new SAXReader(false); reader.setIncludeInternalDTDDeclarations(false); reader.setIncludeExternalDTDDeclarations(false); Document document = reader.read(new StringReader(xmltext)); List nodelist = document.selectNodes(xpath); // Process the elements in the nodelist ArrayList outputList = new ArrayList(); ArrayList outputXmlList = new ArrayList(); String val = null; String xmlVal = null; for (Iterator iter = nodelist.iterator(); iter.hasNext();) { Node element = (Node) iter.next(); xmlVal = element.asXML(); val = element.getStringValue(); if (val != null && !val.equals("")) { outputList.add(val); outputXmlList.add(xmlVal); } } List nodelist=outputList; List nodelistAsXML=outputXmlList; |
extractCount | localworker |
Scriptimport org.dom4j.Document; import org.dom4j.Node; import org.dom4j.io.SAXReader; SAXReader reader = new SAXReader(false); reader.setIncludeInternalDTDDeclarations(false); reader.setIncludeExternalDTDDeclarations(false); Document document = reader.read(new StringReader(xmltext)); List nodelist = document.selectNodes(xpath); // Process the elements in the nodelist ArrayList outputList = new ArrayList(); ArrayList outputXmlList = new ArrayList(); String val = null; String xmlVal = null; for (Iterator iter = nodelist.iterator(); iter.hasNext();) { Node element = (Node) iter.next(); xmlVal = element.asXML(); val = element.getStringValue(); if (val != null && !val.equals("")) { outputList.add(val); outputXmlList.add(xmlVal); } } List nodelist=outputList; List nodelistAsXML=outputXmlList; |
parametersXML_1 | xmlsplitter | |
run_eInfo | wsdl |
Wsdlhttp://eutils.ncbi.nlm.nih.gov/soap/v2.0/eutils.wsdlWsdl Operationrun_eInfo |
run_eSearch_2 | wsdl |
Wsdlhttp://eutils.ncbi.nlm.nih.gov/soap/v2.0/eutils.wsdlWsdl Operationrun_eSearch |
run_eSearch_request | xmlsplitter | |
GENE_RankPhenotypeTerms | workflow | |
max_return_gene | stringconstant |
Value500 |
PATHWAY_RankPhenotypeTerms | workflow | |
split_search_terms_2 | localworker |
ScriptList split = new ArrayList(); if (!string.equals("")) { String regexString = ","; if (regex != void) { regexString = regex; } String[] result = string.split(regexString); for (int i = 0; i < result.length; i++) { split.add(result[i]); } } |
pathway_and_abstract | beanshell |
ScriptString[] split = abstracts.split("\n"); String pathway_name = pathway; Vector nonEmpty = new Vector(); for (int i = 0; i < split.length; i++) { String trimmed = split[i].trim(); nonEmpty.add(trimmed); } String output = ">> " + pathway_name + "\n"; for (int i = 0; i < nonEmpty.size(); i++) { output = output + (String) (nonEmpty.elementAt(i) + "\n"); } |
remove_nulls_3 | beanshell |
ScriptString[] split = input.split("\n"); Vector nonEmpty = new Vector(); for (int i = 0; i < split.length; i++){ if (!(split[i].equals(""))) { nonEmpty.add(split[i].trim()); } } String[] non_empty = new String[nonEmpty.size()]; for (int i = 0; i < non_empty.length; i ++) { non_empty[i] = nonEmpty.elementAt(i); } String output = ""; for (int i = 0; i < non_empty.length; i++) { output = output + (String) (non_empty[i] + "\n"); } |
merge_outputs_2_2 | localworker |
ScriptString seperatorString = "\n"; if (seperator != void) { seperatorString = seperator; } StringBuffer sb = new StringBuffer(); for (Iterator i = stringlist.iterator(); i.hasNext();) { String item = (String) i.next(); sb.append(item); if (i.hasNext()) { sb.append(seperatorString); } } concatenated = sb.toString(); |
regex_2 | stringconstant |
Value\n |
extract_terms | beanshell |
ScriptString[] split = input.split("\n"); Vector nonEmpty = new Vector(); for (int i = 0; i < split.length; i++) { String mytext = split[i].substring(split[i].indexOf(" "), split[i].lastIndexOf(" - ")); nonEmpty.add(mytext); } String output = ""; for (int i = 0; i < nonEmpty.size(); i++) { output = output + (String) (nonEmpty.elementAt(i) + "\n"); } |
add_MeSH_to_string | beanshell |
ScriptString[] split = input.split("\n"); Vector nonEmpty = new Vector(); for (int i = 0; i < split.length; i++) { String trimmed = split[i].trim(); nonEmpty.add(trimmed); } String output = ""; for (int i = 0; i < nonEmpty.size(); i++) { output = output + (String) (nonEmpty.elementAt(i) + " AND \"Metabolic Networks and Pathways\"[MeSH Terms]" + "\n"); } |
Search_PubMed_2 | workflow | |
clean_text_2 | workflow | |
generateTerminology | wsdl |
Wsdlhttp://projects.biotec.tu-dresden.de/DOG4DAG_TAVERNA/services/GoPubMedTermGeneration?wsdlWsdl OperationgenerateTerminology |
generateTerminology_input | xmlsplitter | |
applicationCode_value | stringconstant |
Value20110209_taverna |
Name | Description | Inputs | Outputs |
---|---|---|---|
remove_nulls | input | output | |
gene_and_abstract |
pathway abstracts |
output | |
remove_Nulls_2 | input | output | |
remove_Nulls_4 | input | output | |
concat_abstract_ids |
id abstract_text date_text |
output | |
merge_phenotype_abstracts |
phenotype abstracts |
output | |
extract_abstracts | input |
output output_search |
|
merge_search_and_terms |
terms search_term |
output | |
remove_Nulls_5 | input | output | |
extract_terms_3 | input | output | |
merge_term_count |
term count |
output | |
pathway_and_abstract |
pathway abstracts |
output | |
remove_nulls_3 | input | output | |
extract_terms | input | output | |
add_MeSH_to_string | input | output | |
format_rankings | ranked_terms | title_term_rankings | |
remove_Nulls | input | output | |
stringToBytes | string | bytes | |
concat_abstract_ids |
id abstract_text date_text |
output | |
concat_abstract_ids |
id abstract_text date_text |
output | |
stringToBytes | string | bytes | |
stringToBytes | string | bytes |
Name | Description |
---|---|
gene_abstracts | |
phenotype_abstracts | |
phenotype_terms | |
phenotype_term_counts | |
pubmed_abstract_number | |
gene_cosine | |
gene_term_enrichment_scores | |
pathway_abstracts | |
pathway_cosine_vector_scores | |
pathway_concept_rankings |
Source | Sink |
---|---|
gene_and_abstract:output | remove_nulls:input |
Search_PubMed:concat_data | gene_and_abstract:abstracts |
split_search_terms:split | gene_and_abstract:pathway |
regex:value | split_search_terms:regex |
remove_Nulls_2:output | split_search_terms:string |
clean_text:cleaned_text | merge_outputs_2:stringlist |
split_search_terms:split | Search_PubMed:search_term |
min_date:value | Search_PubMed:min_date |
max_date:value | Search_PubMed:max_date |
max_return_gene:value | Search_PubMed:max_return |
remove_nulls:output | clean_text:input |
split_gene_names:split | Remove_duplicate_genes:stringlist |
gene_names | split_gene_names:string |
regular_expression:value | split_gene_names:regex |
Merge_string_list_to_string:concatenated | remove_Nulls_2:input |
Remove_duplicate_genes:strippedlist | Merge_string_list_to_string:stringlist |
parametersXML_eSearch:output | run_eSearch:parameters |
xpath:value | extractPMID:xpath |
run_eSearch:parameters | extractPMID:xml-text |
merge_phenotype_abstracts:output | remove_Nulls_4:input |
extractPMID:nodelist | concat_abstract_ids:id |
merge_abstracts:concatenated | concat_abstract_ids:abstract_text |
merge_dates:concatenated | concat_abstract_ids:date_text |
merge_abstract_ids:concatenated | merge_phenotype_abstracts:abstracts |
Phenotype_search_term | merge_phenotype_abstracts:phenotype |
Fetch_Abstracts:abstracts | merge_abstracts:stringlist |
max_return_phenotype:value | parametersXML_eSearch:RetMax |
pubmed_database:value | parametersXML_eSearch:db |
max_date:value | parametersXML_eSearch:maxdate |
min_date:value | parametersXML_eSearch:mindate |
Phenotype_search_term | parametersXML_eSearch:term |
concat_abstract_ids:output | merge_abstract_ids:stringlist |
Fetch_Abstracts:pubmed_dates | merge_dates:stringlist |
extractPMID:nodelist | Fetch_Abstracts:pubmed_ids |
remove_Nulls_4:output | clean_text_3:input |
clean_text_3:cleaned_text | Encode_byte_to_base64:bytes |
merge_strings:concatenated | merge_strings_2:stringlist |
clean_text_4:output | extract_abstracts:input |
merge_strings_2:concatenated | merge_search_and_terms:terms |
extract_abstracts:output_search | merge_search_and_terms:search_term |
xpath_2:value | extract_Terms_2:xpath |
generateTerminology:parameters | extract_Terms_2:xml-text |
Remove_duplicate_strings_2:strippedlist | merge_strings:stringlist |
merge_search_and_terms:output | remove_Nulls_5:input |
extract_Terms_2:nodelist | Remove_duplicate_strings_2:stringlist |
regex_3:value | split_by_regex:regex |
extract_abstracts:output | split_by_regex:string |
Encode_byte_to_base64:base64 | clean_text_4:file_direct_data |
remove_Nulls_5:output | extract_terms_3:input |
split_extracted_terms:split | merge_term_count:term |
merge_extracted:concatenated | merge_term_count:count |
extract_terms_3:output | split_extracted_terms:string |
regular_expression_2:value | split_extracted_terms:regex |
extractCount:nodelist | merge_pubmed_count:stringlist |
extractCount_2:nodelist | merge_extracted:stringlist |
merge_term_count:output | merge_list:stringlist |
xpath_3:value | extractCount_2:xpath |
run_eSearch_2:result | extractCount_2:xml-text |
xpath_count:value | extractCount:xpath |
run_eInfo:result | extractCount:xml-text |
pubmed_database_2:value | parametersXML_1:db |
parametersXML_1:output | run_eInfo:request |
run_eSearch_request:output | run_eSearch_2:request |
eSearch_database:value | run_eSearch_request:db |
count:value | run_eSearch_request:rettype |
split_extracted_terms:split | run_eSearch_request:term |
min_date:value | run_eSearch_request:mindate |
max_date:value | run_eSearch_request:maxdate |
merge_pubmed_count:concatenated | GENE_RankPhenotypeTerms:pubmed_abstract_number |
merge_list:concatenated | GENE_RankPhenotypeTerms:phenotype_term_counts |
merge_outputs_2:concatenated | GENE_RankPhenotypeTerms:query_abstracts |
clean_text_3:cleaned_text | GENE_RankPhenotypeTerms:phenotype_abstracts |
remove_Nulls_5:output | GENE_RankPhenotypeTerms:phenotype_terms |
merge_outputs_2_2:concatenated | PATHWAY_RankPhenotypeTerms:query_abstracts |
clean_text_3:cleaned_text | PATHWAY_RankPhenotypeTerms:phenotype_abstracts |
merge_pubmed_count:concatenated | PATHWAY_RankPhenotypeTerms:pubmed_abstract_number |
merge_list:concatenated | PATHWAY_RankPhenotypeTerms:phenotype_term_counts |
remove_Nulls_5:output | PATHWAY_RankPhenotypeTerms:phenotype_terms |
add_MeSH_to_string:output | split_search_terms_2:string |
regex_2:value | split_search_terms_2:regex |
Search_PubMed_2:concat_data | pathway_and_abstract:abstracts |
split_search_terms_2:split | pathway_and_abstract:pathway |
pathway_and_abstract:output | remove_nulls_3:input |
clean_text_2:cleaned_text | merge_outputs_2_2:stringlist |
pathway_descriptions | extract_terms:input |
extract_terms:output | add_MeSH_to_string:input |
split_search_terms_2:split | Search_PubMed_2:Pathway_search_term |
remove_nulls_3:output | clean_text_2:input |
generateTerminology_input:output | generateTerminology:parameters |
split_by_regex:split | generateTerminology_input:texts |
applicationCode_value:value | generateTerminology_input:applicationCode |
merge_outputs_2:concatenated | gene_abstracts |
clean_text_3:cleaned_text | phenotype_abstracts |
remove_Nulls_5:output | phenotype_terms |
merge_list:concatenated | phenotype_term_counts |
merge_pubmed_count:concatenated | pubmed_abstract_number |
GENE_RankPhenotypeTerms:cosine_vector_scores | gene_cosine |
GENE_RankPhenotypeTerms:concept_rankings | gene_term_enrichment_scores |
merge_outputs_2_2:concatenated | pathway_abstracts |
PATHWAY_RankPhenotypeTerms:cosine_vector_scores | pathway_cosine_vector_scores |
PATHWAY_RankPhenotypeTerms:concept_rankings | pathway_concept_rankings |
None
Workflow Type
Version 1 (earliest) (of 2)
- annotation
- |
- concept
- |
- concept profile
- |
- controlled vocabulary
- |
- cosine vector space
- |
- data integration
- |
- data-driven
- |
- dresden
- |
- efetch
- |
- enrichment
- |
- entity recognition
- |
- esearch
- |
- eutils
- |
- gene
- |
- gene identifier
- |
- geneid
- |
- genotype
- |
- kegg
- |
- Kegg Pathways
- |
- KeggID
- |
- literature
- |
- loci
- |
- locus
- |
- medline
- |
- mesh
- |
- pathway
- |
- pathway-driven
- |
- pathways
- |
- phenotype
- |
- pubmed
- |
- qtl
- |
- quanitative
- |
- text mining
- |
- text mining; term extraction; entity recognition
- |
- trait
- |
- triat
- annotation
- |
- concept
- |
- concept profile
- |
- controlled vocabulary
- |
- cosine vector space
- |
- data integration
- |
- data-driven
- |
- dresden
- |
- efetch
- |
- enrichment
- |
- entity recognition
- |
- esearch
- |
- eutils
- |
- gene
- |
- gene identifier
- |
- geneid
- |
- genotype
- |
- kegg
- |
- Kegg Pathways
- |
- KeggID
- |
- literature
- |
- loci
- |
- locus
- |
- medline
- |
- mesh
- |
- pathway
- |
- pathway-driven
- |
- pathways
- |
- phenotype
- |
- pubmed
- |
- qtl
- |
- quanitative
- |
- text mining
- |
- text mining; term extraction; entity recognition
- |
- trait
- |
- triat
Log in to add Tags
Shared with Groups (0)
None
Log in to add to one of your Packs
Statistics
In chronological order:
-
Created by Paul Fisher on Thursday 10 February 2011 16:10:52 (UTC)
Last edited by Paul Fisher on Thursday 10 February 2011 16:15:43 (UTC)
-
Created by Paul Fisher on Friday 18 February 2011 13:45:51 (UTC)
Last edited by Paul Fisher on Friday 18 February 2011 13:47:08 (UTC)
Revision comment:Updated the description and added dates to term counting
Reviews (0)
Other workflows that use similar services (35)
Only the first 2 workflows that use similar services are shown. View all workflows that use these services.
Gene to Pubmed (4)
Created: 2011-02-08 | Last updated: 2011-02-10
Credits: Paul Fisher
Attributions: Cosine vector space Extract Scientific Terms Rank Phenotype Terms Cosine vector space Rank Phenotype Terms Pathway to Pubmed Extract Scientific Terms
PubMed Search (1)
Created: 2011-02-03 | Last updated: 2011-02-03
Credits: Paul Fisher
Comments (0)
No comments yet
Log in to make a comment