similar to: download/retain text file structure with RCurl/getURL()

Displaying 20 results from an estimated 700 matches similar to: "download/retain text file structure with RCurl/getURL()"

2009 Jan 12
2
problems with download.file() from ftp?
Dear list, I am trying to download a text file from an ftp site using download.file(). I used the following code: url <- "ftp://ftp.wcc.nrcs.usda.gov/data/snow/snow_course/table/history/idaho/13e19.txt" dest <- "c:/test/downloadtest.txt" download.file(url, dest) I get this error message, indicating that R is unable to open the connection: trying URL
2010 Sep 16
2
FTP Download
Hi, I have problems downloading complete folders via ftp with R. Single files work fine. I tried Rcurl, but it does not work. This is my code: url = "ftp://disc2.nascom.nasa.gov/data/TRMM/Gridded/Derived_Products/3B42_V6/Daily/2009/" filenames = getURL(url, ftp.use.epsv = FALSE, ftplistonly = TRUE, crlf = TRUE) filenames = paste(url, strsplit(filenames, "\r*\n")[[1]], sep =
2012 Oct 11
1
Problems with getURL (RCurl) to obtain list files of an ftp directory
Dear all, I have a problem with the command 'getURL' from the RCurl package, which I have been using to obtain a ftp directory list from the MOD16 (ET, DSI) products, and then to download them. (part of the script by Tomislav Hengl, spatial-analyst). Instead of the list of files (from ftp), I am getting the complete html code. Anyone knows why this might happen? This are the steps i
2001 Feb 12
2
patches for UnixWare v2.x pty (misc.c,configure.in)
I have attached two patches to the current source code. The first addresses the pty problems with UnixWare 2.x with connecting with SSH2. It sets the sigaction to SA_RESTART. This fixes UnixWare v2.x, but haven't heard any feedback as to effects on other OS'. The first patch is against misc.c. The second patch adds a section "*-*-sysv4.2uw2*" to configure to set the
2011 Jul 01
13
For help in R coding
Dear all, I am doing a project on variant calling using R.I am working on pileup file.There are 10 columns in my data frame and I want to count the number of A,C,G and T in each row for column 9.example of column 9 is given below- .a,g,, .t,t,, .,c,c, .,a,,, .,t,t,t .c,,g,^!. .g,ggg.^!, .$,,,,,.,
2012 Mar 19
2
fitted values with locfit
Dear memberships, I'm trying to estimate the following multivariate local regression model using the "locfit" package: BMI=m1(RCC)+m2(WCC) where (m1) and (m2) are unknown smooth functions. My problem is that once I get the regression done I cannot get the fitted values of each of this smooth functions (m1) and (m2). What I write is the following library(locfit) data(ais)
2001 Jan 30
3
dsa_verify signature incorrect
I am building version 2.3.0p1 of openssh on a UnixWare 2.03 system and am unable to connect with SSH 2. The error I get is: debug: len 55 datafellows 0 debug: dsa_verify: signature incorrect dsa_verify failed for server_host_key The build environment is as follows: gcc 2.95.1 openssl-0.9.6-beta2 I've looked through the archives and found similar problems related to version
2004 Aug 06
1
Icecast server YP Woes... take 2
I am having problem with icecast2/winamp oddcast dsp/darkice being listed on dir.xiph.org and dir.oddsock.org. Now I've scanned the mailing list and found that there was a previous issue, however if I use icecast on Windows XP it works fine, but not from Fedora Core 1. I've made sure to check my config files multiple times and looked at the error logs but I still can't figure
2013 Feb 21
4
Getting htmlParse to work with Hebrew? (on windows)
Hello dear R-help mailing list. Looks like the same issue in Russian: library(RCurl) library(XML) u = " http://www.cian.ru/cat.php?deal_type=2&obl_id=1&room1=1" a = getURL(u) a # Here - the Russian is fine. a2 <- htmlParse(a) a2 # Here it is a mess... None of these seem to fix it: htmlParse(a, encoding = "windows-1251") htmlParse(a, encoding =
2009 Jan 26
2
RCurl unable to download a particular web page -- what is so special about this web page?
Dear R-help, There seems to be a web page I am unable to download using RCurl. I don't understand why it won't download: > library(RCurl) > my.url <- "http://www.nytimes.com/2009/01/07/technology/business-computing/07program.html?_r=2" > getURL(my.url) [1] "" Other web pages are ok to download but this is the first time I have been unable to download a
2004 Aug 06
2
Icecast server YP Woes... take 2
can you provide a dump of the conversation between your icecast server and the YP ? Use something like tcpdump or ethereal...In the dump, look for the URL call to yp-cgi and post what is being sent... oddsock <p>What I saw of the output of tcpdump didn't show any mention of the yp url. The communications seem to run only between the client and the server and nowhere else. This is the
2004 Aug 06
6
Runaway icecast again (more YP issues)
Hi: The sysadmin on the system where the icecast server I more or less administer runs sent me an E-mail today saying that icecast had pushed up the system load to over 8 and that he had to kill it off. Some discussion and examination of the log files suggest that there's a problem with removing streams from the YP server. The main source on our server is on a machine that's used for
2004 Aug 06
6
Runaway icecast again (more YP issues)
Hi: The sysadmin on the system where the icecast server I more or less administer runs sent me an E-mail today saying that icecast had pushed up the system load to over 8 and that he had to kill it off. Some discussion and examination of the log files suggest that there's a problem with removing streams from the YP server. The main source on our server is on a machine that's used for
2010 Jul 21
1
Command that is conditional upon file retrieval: is it possible?
Hi all, I'm currently working on an R program where I have to access an FTP server to download some of the data I need. However, the people who post up the files I access are at times inconsistent with regards to time posted, if they post at all, etc.... Here's some of the code I use: library(RCurl) url1 = paste("ftp://user:password at a.great.website.com/",
2004 Aug 06
2
YP Server Woes
I'm having a problem getting my stream published on dir.xiph.org using icecast2/ices2-beta4 The output in my log is: [snip] DBUG geturl/curl_print_header_result SID -> () DBUG geturl/curl_print_header_result Message -> () DBUG geturl/curl_print_header_result Touch Freq -> (0) DBUG geturl/curl_print_header_result Response -> (0) EROR yp/yp_submit_url Got a NAK from yp_add(Unknown)
2009 Sep 17
1
RCurl and Google Scholar's EndNote references
Hi! I've performed a Google Scholar Search using a query, let's say "Frank Harrell", and parsed the links to the EndNote references from the resulting HTML code. Now I'd like to download all the references automatically. For this, I have tried to use RCurl, but I can't seem to get it working: I always get error code "403 Forbidden" from the web server.
2010 Nov 14
1
RCurl and cookies in POST requests
Hello. I know that it's usually possible to write cookies to a cookie file by removing the curl handle and doing a gc() call. I can do this with getURL(), but I just can't obtain the same results with postForm(). If I use: curlHandle <- getCurlHandle(cookiefile=FILE, cookiejar=FILE) and then do: getURL(http://example.com/script.cgi, curl=curlHandle) rm(curlHandle) gc() it's
2010 Aug 17
1
navel-gazing
month <- "2010-August" list <- "r-help" ##list <- "r-sig-ecology" ##list <- "r-sig-mixed-models" ## month <- "2010q3" n <- 50 baseurl <- "https://stat.ethz.ch/pipermail/" library(RCurl) z <- getURL(paste(baseurl,list,"/",month,"/author.html",sep="")) zz <-
2016 Apr 15
1
Need Help to solve an Error in R
Hi Keep your answers to R help (others can help you too) From the warning message it seems to me possible that function ReadExi needs to write something to the working directory. As I said I am not an expert in this package, but commands from help R> make.gal.env(galname='galenv', gal.path='Exiqon') R> ebatch <- ReadExi(galname='galenv',
2013 Jul 23
2
downloading web content
Hello, I am trying to use R to download a bunch of .csv files such as: http://biocache.ala.org.au/ws/occurrences/download?q=Banksia+ericifolia I have tried the following and neither work: a<- getURL(" http://biocache.ala.org.au/ws/occurrences/download?q=Banksia+ericifolia") Error in curlPerform(curl = curl, .opts = opts, .encoding = .encoding) : embedded nul in string: and