Displaying 20 results from an estimated 1000 matches similar to: ""Special" characters in URI"
2005 May 08
2
Extract just some fields from XML
Hello!
I am trying to get specific fields from an XML document and I am totally
puzzled. I hope someone can help me.
# URL
URL<-"http://eutils.ncbi.nlm.nih.gov/entrez/eutils/efetch.fcgi?db=pubmed&id=11877539,11822933,11871444&retmode=xml&rettype=citation"
# download a XML file
tmp <- xmlTreeParse(URL, isURL = TRUE)
tmp <- xmlRoot(tmp)
Now I want to extract only
2007 Dec 14
6
Analyzing Publications from Pubmed via XML
I would like to track in which journals articles about a particular disease
are being published. Creating a pubmed search is trivial. The search
provides data but obviously not as an R dataframe. I can get the search to
export the data as an xml feed and the xml package seems to be able to read
it.
xmlTreeParse("
2005 May 10
0
Fwd: Extract just some fields from XML]
Duncan, you are a king!
Thanks a lot for this cookie. It really helped me. Thanks for the code
as well as detailed explanation at the end.
>Hi Gregor.
>
>Here is a function that will collect all of the nodes in the
>XML document whose names are in the vector elementNames
>
>getElements =
>function(elementNames)
>{
> els = list()
>
> startElement = function(node,
2012 Dec 11
1
query multiple terms in PubMed abstract
Hi:
I am trying to search PubMed abstracts which contains BOTH two terms:
COL4A1 AND Ocular. I am using the following code:
url= "http://eutils.ncbi.nlm.nih.gov/entrez/eutils/esearch.fcgi?"
search = paste(url,
"db=pubmed&term=COL4A1+AND+Ocular[abstract]&retmax=300", sep="")
docId <- xmlTreeParse(getURL(paste(url, search, sep="")),
2012 Dec 27
1
Conjunction and disjunction in pubmed query
Hi:
I am trying to query pubmed abstracts using the following syntax:
url= "http://eutils.ncbi.nlm.nih.gov/entrez/eutils/esearch.fcgi?"
search = paste(url, "db=pubmed&term=", queryTerm1, "+AND+",
queryTerm2,"+OR+",queryTerm3, "+OR+", queryTerm4,
"[abstract]&retmax=100&usehistory=y", sep="")
docId <-
2005 Jun 14
1
protection stack overflow??
Hi dear Rers,
I am using SSOAP package to access SOAP service at NCBI.
I followed the example code in SSOAP but failed.
> z <- .SOAP("http://www.ncbi.nlm.nih.gov/entrez/eutils/soap/soap_adapter.cgi", method="run_eInfo", db="pubmed", action = I("einfo"))
Error: protect(): protection stack overflow
what's wrong?
Thanks very much.
Regards
2005 Jun 14
0
question about SSOAP
Dear R folks:
I am trying to use SSOAP (version 0.2-2) package in R (version
2.1.0,linux) to access SOAP service on NCBI
(http://www.ncbi.nlm.nih.gov)
its WSDL file is at http://www.ncbi.nlm.nih.gov/entrez/eutils/soap/eutils.wsdl
but some errors occured:
> ncbi <- processWSDL("http://www.ncbi.nlm.nih.gov/entrez/eutils/soap/eutils.wsdl")
> ff <-
2008 Feb 18
2
Huge number
Hi,
I'm trying to calculate p-value to findout definitely expressed genes
compare A to B situation.
I got this data(this is a part of data) from whole organism , and each
number means each expression values (that means, we could think 'a' gene
is 13 in A situation, and it turns 30 in B situation)
To findout probability, I'm going to use Audic - Claverie Method. ( The
significance
2017 Jun 09
1
efetch result not in character format
Hi,
I want to use reutils to obtain the accession numbers of a query search in
character format. When I use efetch, the accession number isn't in a
character format, and I'm not sure if the number is accurate, because I get
the error:
Error in file.exists(destfile) : object 'destfile' not found
This is what I tried:
UIDs<-esearch( "Methylation" )
accession_numbers
2011 Apr 07
1
Two questions about metacharacter in regexprs and function return
for the script, please kindly see the script below. At line 10 and line 13,
my problems occurs.
The first one is I try to retrieve the gene official name from a column of a
table. The pattern of official name is something starting with gene_name.
For detail problems, please see the according lines.
Any suggestions are appreciated
example of matching source (extract the Nnat, sometime it would
2004 Sep 29
2
RSXML - Parsing XML Documents on Internet
R Users -
I asked about this a few months ago and never did quite figure it out, so with more information, allow me to try again.
If I use the following code:
library(xml)
xmlTreeParse("http://home.comcast.net/~larsenmtl/xmlTestDoc.xml", isURL = TRUE)
I receive this error:
Error in xmlTreeParse("http://home.comcast.net/~larsenmtl/xmlTestDoc.xml"", :
error in
2011 Jul 05
2
Stuck ...can't get sapply and xmlTreeParse working
Can't seem to get the code below working. It gets stuck on line 24 inside the
function hm; comments show the line in question. The function hm is called
by sapply and is at the bottom of the code. Other stuff above line 24 works
correctly including the first couple of lines of the function hm. Should I
be using a different apply function or am I doing something wrong with
xmlTreeParse ?
2011 Jun 07
1
XML segfault on some architectures
Hi,
I found an architecture-specific segfault problem with the XML package. I originally found the problem using the parseKGML2Graph function in the Bioconductor KEGGgraph package, but as far as I can tell the underlying issue seems to be with the xmlTreeParse which is called by parseKGML2Graph.
I'm trying this piece of code, from the xmlTreeParse help page:
library(XML)
fileName <-
2008 Mar 18
1
problem with merge
I have used merge regularly and thought I understood how it worked, but
I must not. I have two dataframes with identical colnames from two
different experiments, TL01 and LC01. Each dataframe has a column named
"Entrez.Gene", which I have converted to "as.character" just to make
sure merge is not looking at factor levels. Because I have done some
filtering, the Entrez.Gene
2008 Jul 02
1
Removing or overwriting an XML node
Hi,
I have an existing XML document on disk, which I'd like
to use as a template, and exchange a subnode with my own
newly created subtree:
<?xml version="1.0"?>
<Duncan>
<name a="1" b="xyz">
<first>Duncan</first>
<last>Temple Lang</last>
</name>
</Duncan>
created by e.g.
? library(XML)
2007 Nov 14
1
reading tables from url
I'm trying to read some web tables directly into R. These are both
genome sequencing projects (eukaryotes and metagenomes) from NCBI and
look very similar; however, only the first one works.
http://www.ncbi.nlm.nih.gov/genomes/leuks.cgi
http://www.ncbi.nlm.nih.gov/genomes/lenvs.cgi
I added ?dump=selected to the end of the url string to get a tab-
delimited file (which is what happens
2010 Oct 22
1
cannot connect to an FTP server with long HELLO message
Hi,
Trying to access files on the ftp server at ftp.ncbi.nih.gov
will either give a time out or sometimes even a segfault on Linux.
The 2 following methods give the same results:
f <-
url("ftp://ftp.ncbi.nih.gov/pub/geo/DATA/SOFT/GDS/GDS10.soft.gz", open="r")
download.file("ftp://ftp.ncbi.nih.gov/pub/geo/DATA/SOFT/GDS/GDS10.soft.gz",
destfile=tempfile())
2005 Jul 19
1
Minor "bug" in source()
For R v2.1.1 patched and R v2.2.0 devel:
Calling source(file, chdir=TRUE) with is.character(file) != TRUE, that
is, with 'file' as a connection, will generate an error. Example:
> file <- textConnection("cat('Hello world\n')")
> source(file, chdir=TRUE)
Error in source(file, chdir = TRUE) : Object "ofile" not found
Of course, it does not make
2005 Feb 07
2
Programming/scripting with "expressions - variables"
Hello to Rusers!
I am puzzled with R and I really do not know where to look
in for my problem. I am moving from SAS and I have difficulties
in translating SAS to R world. I hope I will get some hints
or pointers so I can study from there on.
I would like to do something like this. In SAS I can write
a macro as example bellow, which is afcourse a silly one but
shows what I don't know how
2009 Nov 19
1
problem post request with RCurl
Hi, I am trying to use a CGI service (Pubchem PUG) via RCurl and am
running into a problem where the data must be supplied via POST - but
I don't know the keyword for the argument.
The data to be sent is an XML fragment. I can do this via the command
line using curl: I save the XML string to a file called query.xml and
then do
curl -d @query.xml