Displaying 20 results from an estimated 3000 matches similar to: "ftp fetch using RCurl?"
2012 May 28
1
Rcurl, postForm()
Dear colleagues,
Could I get some assistance using postForm() to scrape the business names and addresses at this website:
http://www.brantford.ca/business/LocalBusinessCommunity/Pages/BusinessDirectorySearch.aspx
I've read through (http://www.omegahat.org/RCurl/RCurlJSS.pdf) and scoured the web for tutorials, but I can't crack it. I'm aware that this is probably a pretty basic
2012 Jun 07
1
How to set cookies in RCurl
Hi,
I am trying to access a website and read its content. The website is a
restricted access website that I access through a proxy server (which
therefore requires me to enable cookies). I have problems in allowing Rcurl
to receive and send cookies.
The following lines give me:
library(RCurl)
library(XML)
url <- "http://www.theurl.com"
content <- readHTMLTable(url)
content
2009 Sep 17
1
RCurl and Google Scholar's EndNote references
Hi!
I've performed a Google Scholar Search using a query, let's say "Frank
Harrell", and parsed the links to the EndNote references from the resulting
HTML code. Now I'd like to download all the references automatically. For
this, I have tried to use RCurl, but I can't seem to get it working: I
always get error code "403 Forbidden" from the web server.
2010 Nov 04
3
postForm() in RCurl and library RHTMLForms
Hi RUsers,
Suppose I want to see the data on the website
url <- "http://www.nseindia.com/content/indices/ind_histvalues.htm"
for the index "S&P CNX NIFTY" for
dates "FromDate"="01-11-2010","ToDate"="02-11-2010"
then read the html table from the page using readHTMLtable()
I am using this code
webpage <-
2011 Jan 06
8
Accessing data via url
# Can anyone suggest why this works
datafilename <- "http://personality-project.org/r/datasets/maps.mixx.epi.bfi.data"
person.data <- read.table(datafilename,header=TRUE)
# but this does not?
dd <- "https://sites.google.com/site/jrkrideau/home/general-stores/trees.txt"
treedata <- read.table(dd, header=TRUE)
2010 Nov 10
3
RGoogleDocs stopped working
Hello,
Some code using RGoogleDocs, which had been working smoothly since the
summer, just stopped working. I know that it worked on November 3rd, but it
doesn't work today. I've confirmed that the login and password still work
when I log in manually. I've confirmed that the URL gives the same error
when I paste it into Firefox. I don't know enough about this web service to
figure
2012 Sep 26
1
RCURL ftp upload - ASCII or Binary type?
I'm trying to upload a file using RCURL:s ftpUpload() to a ftp-server using the following command:
> ftpUpload("'VERY.ODD.FILE.NAME(+1)'",to="ftp://x.x.x.x/","' VERY.ODD.FILE.NAME(+1)'",userpwd="USER:PASSWORD")
OK
0
The file I'm trying to upload is a very simple text-file but with a bit weird filename. Note the ' on each
2009 Oct 15
1
Removing Embedded Null characters from text/html
Hi,
I'm trying to download some data from the web and am running into
problems with 'embedded null' characters. These seem to indicate to R
that it should stop processing the page so I'd like to remove them.
I've been looking around and can't seem to identify exactly what the
character is and consequently how to remove it.
# THE CODE WORKS ON THIS PAGE
library(RCurl)
2012 May 14
3
Scraping a web page.
Folks,
I want to scrape a series of web-page sources for strings like the following:
"/en/Ships/A-8605507.html"
"/en/Ships/Aalborg-8122830.html"
which appear in an href inside an <a> tag inside a <div> tag inside a table.
In fact all I want is the (exactly) 7-digit number before ".html".
The good news is that as far as I can tell the the <a>
2007 Mar 21
2
problem with RCurl install on Unix
I am having trouble getting an install of RCurl to work properly on a
Unix server. The steps I have taken are:
1. installed cUrl from source without difficulty
2. installed RCurl from source using the command
~/R_HOME/R-devel/bin/R CMD INSTALL -l ~/R_HOME/R-devel/library
~/RCurl_0.8-0.tar.gz I received no errors during this install
3. when I go back to R and require(RCurl), I get
>
2007 Mar 21
2
problem with RCurl install on Unix
I am having trouble getting an install of RCurl to work properly on a
Unix server. The steps I have taken are:
1. installed cUrl from source without difficulty
2. installed RCurl from source using the command
~/R_HOME/R-devel/bin/R CMD INSTALL -l ~/R_HOME/R-devel/library
~/RCurl_0.8-0.tar.gz I received no errors during this install
3. when I go back to R and require(RCurl), I get
>
2008 Dec 01
1
[BioC] Rcurl 0.8-1 update for bioconductor 2.7
Hi Patrick,
Greetings from !(sunny) Pittsburgh.
What's the scoop on RCurl on windows (XP)?
I've tried to install RCurl_0.92-0.zip and RCurl_0.9-3.zip,
with both R 2.7.2 and R 2.8.0 from the RGUI (utils:::menuInstallLocal),
and get the error
"Windows binary packages in zipfiles are not supported".
which (according to google's one and only hit) comes from a perl script.
2008 Jul 25
1
Installation error for RCurl in Redhat enterrpise 5
I am getting the following error while trying to install the RCurl library. I have checked that the curl and the libcurl.so.3
is already installed in the /usr/bin
> install.packages("RCurl")
--- Please select a CRAN mirror for use in this session ---
Loading Tcl/Tk interface ... done
trying URL 'http://cran.hostingzero.net/src/contrib/RCurl_0.9-3.tar.gz'
Content type
2012 Oct 11
1
Problems with getURL (RCurl) to obtain list files of an ftp directory
Dear all,
I have a problem with the command 'getURL' from the RCurl package, which I
have been using to obtain a ftp directory list from the MOD16 (ET, DSI)
products, and then to download them. (part of the script by Tomislav
Hengl, spatial-analyst). Instead of the list of files (from ftp), I am
getting the complete html code. Anyone knows why this might happen?
This are the steps i
2011 Jun 06
1
RCurl and kerberos
Dear list,
I would like to call a Kerberos-authenticated web-service from within R.
Curl can do it:
$ curl --negotiate -u : "http://my.web.service/"
so I would expect that RCurl also has the capability, but I have not been able to find the correct options to set.
listCurlOptions() does not return anything with negotiate, and searching the source of RCurl, the only thing I found was
2014 Jan 02
2
Installing RCurl -
Dear all,
I am trying to install RCurl (because I want to install devtools) and to do so I've been informed that I must install one of the packages
libcurl4-openssl-dev
libcurl4-nss-dev
No matter which one I install I get the following error from R:
* installing *source* package ‘RCurl’ ...
** package ‘RCurl’ successfully unpacked and MD5 sums checked
checking for curl-config...
2011 Apr 03
1
problem in install RCurl in R (Ubuntu Linux)
I have some problem in running R-cran's Demography package.
The hmd.mx function need Rcurl. I tried to install RCurl, but meet the following error:
*********************************************************************
...
* installing *source* package ?RCurl? ...
checking for curl-config... no
Cannot find curl-config
ERROR: configuration failed for package ?RCurl?
* removing
2012 Sep 19
1
scraping with session cookies
Hi, I am starting coding in r and one of the things that i want to do is to
scrape some data from the web.
The problem that I am having is that I cannot get passed the disclaimer
page (which produces a session cookie). I have been able to collect some
ideas and combine them in the code below but I dont get passed the
disclaimer page.
I am trying to agree the disclaimer with the postForm and write
2010 Dec 03
1
Problem installing RCurl
I have 64-bit R 2 12 0 installed on Solaris 10 of Sun Sparc. When I tried to install RCurl, it failed with the following lines,
...............
Version has CURLOPT_SSL_SESSIONID_CACHE
libcurl version: libcurl 7.19.6
configure: creating ./config.status
config.status: creating src/Makevars
** libs
cc -xc99 -m64 -xarch=sparcvis2 -I/apps/sparcv9/R-2.12.0/lib/R/include -I/opt/csw/include
2009 Feb 18
1
R as a web scraping tool using RCurl
Hi List,
I am trying to leverage my knowledge of R in trying to use it for tasks that
may not make R the best choice for these tasks.
I wish to automate a web scraping task, which requires a multi-step
procedure:
1) log in to a website
2) Go to a particular page
3) From the drop down menu, click on a particular link
4) From the tabulated data presented, choose relevant information based on a