similar to: Strange behaviour to download zip file using R

Displaying 20 results from an estimated 2000 matches similar to: "Strange behaviour to download zip file using R"

2010 Nov 01
1
File Downloading Problem
Dear Group, My code stopped working ... used to work till last week! sURL <- "http://www.nseindia.com/content/historical/EQUITIES/2010/NOV/cm01NOV2010bha v.csv.zip" > download.file(sURL,"test.zip") trying URL 'http://www.nseindia.com/content/historical/EQUITIES/2010/NOV/cm01NOV2010bha v.csv.zip' Error in download.file(sURL, "test.zip") : cannot
2018 May 13
2
is there any method to defer the execution of code in r?
dear members, I have created the following function to read a csv file from a given URL: function(){ s <- 1; #first get the bhav copy today <- c();ty <- c();tm <- c();tmu <- c();td <- c(); # get the URL first today <- Sys.Date()
2018 May 13
0
is there any method to defer the execution of code in r?
Hello, This is cross posted from StackOverflow: https://stackoverflow.com/questions/50314015/is-there-any-method-to-defer-the-execution-of-code-in-r Cross posting is discouraged in r-help. Rui Barradas On 5/13/2018 8:59 AM, akshay kulkarni wrote: > dear members, > > I have created the following function to read a csv file from a given URL: > > function(){ >
2018 May 13
2
is there any method to defer the execution of code in r?
I am puzzled by the use of the term "cross-posted" here... I don't see the OP or their question or any similar words from the question involved the the given link, though that link seems worth bringing it to the OP's attention. But the function given in the question seems to have other problems: A) The download.file function call puts its result in a different place than the
2018 May 13
0
is there any method to defer the execution of code in r?
Hello, I don't understand. It *is* the same question. Same code, same words. And same 'AKSHAY M KULKARNI' (the OP here) and 'AKshayKulkarni' (SO). Exactly the same. Rui Barradas On 5/13/2018 2:07 PM, Jeff Newmiller wrote: > I am puzzled by the use of the term "cross-posted" here... I don't see the OP or their question or any similar words from the
2012 Oct 15
2
Download a file from url
Hi Download the xls file from this link http://www.nseindia.com/content/fo/fii_stats_12-Oct-2012.xls <http://www.nseindia.com/content/fo/fii_stats_12-Oct-2012.xls> I tried this --------------------------------------------------------------- library(XLConnect) url = "http://www.nseindia.com/content/fo/fii_stats_12-Oct-2012.xls" local.xls.file = tempfile() download.file(url,
2007 Oct 16
1
try / tryCatch for download.file( ) within a for loop when URL does not exist
I am trying to download a bunch of files from a server, for which I am using download.file( ) within a for loop. The script is working fine except until download.file hits a URL which has no file, at which point it exits. I want to change this behavior to simple log the failure and maintain state within the for loop and iterate to next. I read about try / tryCatch but am having trouble
2010 Nov 04
3
postForm() in RCurl and library RHTMLForms
Hi RUsers, Suppose I want to see the data on the website url <- "http://www.nseindia.com/content/indices/ind_histvalues.htm" for the index "S&P CNX NIFTY" for dates "FromDate"="01-11-2010","ToDate"="02-11-2010" then read the html table from the page using readHTMLtable() I am using this code webpage <-
2012 Oct 17
0
postForm() in RCurl and library RHTMLForms
Hi R Users, I want to get the data from the url given from 10/09/2012 to 15/10/2012. I don't know how to pass the parameters . ....................................................................................................................................... library(RHTMLForms) > > ff = getHTMLFormDescription("
2012 Dec 02
1
postForm() in RCurl and library RHTMLForms
Hi RUsers, Suppose I want to see the data on the website url <- "http://www.nseindia.com/content/indices/ind_histvalues.htm" for the index "S&P CNX NIFTY" for dates "FromDate"="01-11-2010","ToDate"="02-11-2010" then read the html table from the page using readHTMLtable() I am using this code webpage <-
2018 May 13
3
is there any method to defer the execution of code in r?
Not when I click on that link. On May 13, 2018 7:37:50 AM PDT, Rui Barradas <ruipbarradas at sapo.pt> wrote: >Hello, > >I don't understand. > >It *is* the same question. Same code, same words. And same 'AKSHAY M >KULKARNI' (the OP here) and 'AKshayKulkarni' (SO). > >Exactly the same. > >Rui Barradas > >On 5/13/2018 2:07 PM, Jeff
2008 Oct 07
1
FW: Reading Data
Rahul Agarwal Analyst Equities Quantitative Research UBS_ISC, Hyderabad On Net: 19 533 6363 hi let me explain you the problem we have a database which is in this format Stocks 30-Jan-08 28-Feb-08 31-Mar-08 30-Apr-08 a 1.00 3.00 7.00 3.00 b 2.00 4.00 4.00 7.00 c 3.00 8.00 655.00 3.00 d 4.00 23.00 4.00 5.00 e 5.00 78.00 6.00 5.00 and we have a query
2008 Nov 19
2
Oja median
Hi Roger, As we know that The Oja median has (finite) breakdown point 2/n, i.e., is not robust in any reasonable sense, and is quite expensive to compute, so do we have some better methodology to compute multivariate median Rahul Agarwal Analyst Equities Quantitative Research UBS_ISC, Hyderabad On Net: 19 533 6363 [[alternative HTML version deleted]]
2007 Oct 15
1
String concatenation, File Path Handling to pass to download.file( ) [backslash in DOS paths]
Gabor, Thanks much. Your solution is elegant. My overall scheme is to take present date, and check whether it is a weekend, if not, then create a string based on the date, to concatenate into a url link for download.file( ). The files I need to download have a part which is in the format: mmddyy. I am working to make myself a system to connect to exchanges, and download end of day files from
2011 Aug 23
1
Testing Specific Hypothesis
Hi All! I am interested in testing whether the means for the data I am investigating are equal to a specific value - let's say 0.01. I have already run a one-way ANOVA and know that the differences in the means are not significant, so now I want to know what values the means take on. "otestme" is the data I am working with (it would be hard for me to get into a form that would be
2010 Jun 05
1
How to get the closing price from the the GOOGLE FINANCE site for NSEINDIA stocks
Sir, How to get the closing price from this link http://www.google.com/finance/historical?q=NSE:RCOM I installed quantmod getSymbols('NSE:RCOM',src='google') gives me this error********************** Error in download.file(paste(google.URL, "q=", Symbols.name, "&startdate=", : cannot open URL
2008 Sep 18
3
Oja median
Hi, Can we get the code for calculating Oja median for multivariate data Thanks and Regards Rahul Agarwal Analyst Equities Quantitative Research UBS_ISC, Hyderabad On Net: 19 533 6363 [[alternative HTML version deleted]]
2011 Dec 23
1
Help creating a symmetric matrix?
Hi, I am trying to work with the output of the MINE analysis routine found at http://www.exploredata.net Specifically, I am trying to read the results into a matrix (ideally an n x n x 6 matrix, but I'll settle right now for getting one column into a matrix.) The problem I have is not knowing how to take what amounts to being one half of a symmetric matrix - excluding the diagonal -
2011 Aug 20
1
'install package(s) from local zip files': what is the syntax for that?
Dear all, if I want to install a package in windows system then generally I use the 'install package(s) from local zip files' from the 'package' menu. However I am interested to know that whether there is any syntax which I can use in the R console instead. I have tried with install.packages("__package_name", repos = 'f:/"), however could not. Any suggestion
2006 Jul 06
1
which data structure for a set of time series ?
Hello, I'm a R newcomer and I'm wondering the kind of data structure that would best fit to my problem: my data are equities (stocks) : so I have a time serie (say 1 year of weekly data), and a bunch of qualitative + quantitative variables : the sector of the stock (biotech/finance...), the geographical region, the name, ISIN code, P/E ratios, whatever... The data.frame is perfect for the