similar to: load windows file in linux

Displaying 20 results from an estimated 3000 matches similar to: "load windows file in linux"

2005 Jun 30
2
ts.plot data labels
Dear R users, I am trying to put labels on the data points on a ts.plot. I have tried: ts.plot(df.ts, gpars=list(xy.labels=colnames(df.ts), xlab="group", ylab="level")) which plots fine but gives no labels xy.labels=TRUE also plots without labels I also tried using text() but got an error Error in xy.coords(x, y, recycle = TRUE) : Argument "x" is
2005 Aug 12
1
access assigned objects
Dear list, I am trying to write a function that will make a matrix of each row of a data frame (4 columns), calculate a fisher.test on each resulting matrix and assign a vector of the p-values. I have gotten through making the matrices, but cannot calculate the fisher.test. fm<-function(x) {dfrow<-nrow(x) mm <- vector("list", dfrow) for(i in 1:dfrow) mm[[i]] <-
2007 Oct 10
2
slow load() in R2.6.0
I'm encountering excruciatingly slow load times for character vectors in R 2.6.0-- up to 30sec for a 15K file that contains a no-attributes character vector of length ~1e4 and object size ~0.5MB. In R 2.5.1, repeated loads of the same set of files are near-instantaneous. The problem is proving tricky to reproduce consistently from scratch, so I have attached the 3 files used in the examples
2006 Oct 06
2
Extreme slowdown with named vectors. A bug?
Tried the following with R --vanilla on the Rv2.4.0 release (see details at the end). I think the script and its comments speaks for itself, but the outcome is certainly not wanted. for (n in 58950:58970) { cat("n=", n, "\n", sep=""); # Clean up first rm(names, x, y); gc(); # Create a named vector of length n # Try with format "%5d" and it
2010 Oct 05
2
[LLVMdev] gold-plugin build errors
I tried to build the gold plugin and receive the errors posted below. I checked out the gold plugin using "cvs -z 9 -d :pserver:anoncvs at sourceware.org:/cvs/src co src," as documented on llvm.org. Is that version bleeding edge and perhaps not stable? The first errors have to do with the libtool version. I have libtool 2.2.6 on my system, which is what the README-maintainer-mode
2007 Aug 11
1
LDA and RDA: different training errors
Hello I try to fit a LDA and RDA model to the same data, which has two classes. The problem now is that the training errors of the LDA model and the training error of the RDA model with alpha=0 are not the same. In my understanding this should be the case. Am I wrong? Can someone explain what the reason for this difference could be? Here my code: LDA model: =========== % x is a dataframe tmp =
2008 Aug 22
1
save() should not overwrite a file if an error occurs (PR#12583)
If save() fails because an object is not found, it should not overwrite an existing file. > a <- 1:9 > save(a, file = "a.rda") > rm(a) > load("a.rda") > a [1] 1 2 3 4 5 6 7 8 9 > rm(a) > save(a, file = "a.rda") Error in save(a, file = "a.rda") : object 'a' not found > load("a.rda") Error in
2006 Mar 24
2
[PATCH] qemu pcnet emulation fixes
The attached patch to the qemu emulation of the pcnet hardware fixes several problems. It will now only read and write a transmit or receive descriptor once. It will correctly handle transmitting frames with more than two fragments. It will discard oversize frames instead of corrupting memory. I have tested all the changes I have made and even seen an improvement in receive performance from
2007 Feb 27
2
RDA and trend surface regression
Dear all, I'm performing RDA on plant presence/absence data, constrained by geographical locations. I'd like to constrain the RDA by the "extended matrix of geographical coordinates" -ie the matrix of geographical coordinates completed by adding all terms of a cubic trend surface regression- . This is the command I use (package vegan): >rda(Helling ~
2002 Feb 22
2
R gnome produces errors?: a clue!
I've got something that might be a clue to understand the strange behaviour of lda in R --gui="gnome": in R --vanilla: library(MASS) load("mod23puriflda.rda") > a <- lda.default(x=mod23puriflda[,2:5],grouping=mod23puriflda[,6],CV=F) works fine. I save lda.default as a local rda file: > milda.nognome <- lda.default >
2009 Feb 12
1
Different labels for subsets of points in a PCA or RDA biplot
I've tried a few things both with prcomp(), and rda() and its friends in vegan (including biplot.rda and ordiplot), but can't find a solution. I'd like to associate subsets of the points in a resulting biplot ("sites" in the rda object) with different plotting colors/text styles to emphasize certain sets of points. I can't figure out how to keep the arrows (for
2012 May 16
1
Merging multiple data sets
Hello R user, I have four data sets in dir "D:/Bharat Warule/Rdata_file" which are output_data_prod_1.rda, output_data_prod_2.rda, output_data_prod_3.rda, output_data_prod_4.rda. Each data set is huge size like number of rows 343297 and columns are near to 50. For example: x1 <- data.frame(x11=c(1,2,3,4,5),x112=c(10,10,10,10,10)) x2 <-
2013 Apr 24
1
RDA permutest envfit
Dear all, I did a RDA and when I looked to the signification of the test with permutest, the output was non-significant. But when I used the envfit function, some of the vectors are significant. All the test's conditions are respected. What it means? Is it an error in the script? Commands and output: > permutest(rda.ind, perm=999, first=TRUE) Permutation test for rda  Call:
2002 Jan 12
1
R-1.4.0 delay()
R-1.4.0 seems to have changed the way delay() works, or perhaps it is the way library() works that nullifies the advantages of using delay() in a package. It seems that the process of loading a package evaluates each object provided by that package, and so in particular, the delay()ed object is evaluated and brought into memory. For example generating a test example, where test/R/functions.s is:
2013 Mar 23
1
sysdata.rda vs. rda files in data directory
Dear developeRs, my package FrF2.catlg128 holds large catalogues and is supposed to gain additional ones. All the catalogues are intended for the user. So far, the catalogues were stored in the data directory, and LazyData was "no". I understand that this is not considered wise any more (if it ever was), so that I want to change to LazyData "yes" with the next release
2017 Sep 12
2
Load R data files
Dear All: It was saved, but there was a space somewhere. So it works for me now. I do have another similar problem. I saved an R data file save(datahs0csv,file=" F:\Fall_2017\5-STA574\2-Notes\1-R\1-R_new\chapter4-Entering_Data/datahs0csv2 .rda") *The new R data file "*datahs0csv2.rda*" is in the directory.* I tried to load the file "" to R, but I got an error
2012 Aug 16
1
Compressing data for package builds
Hi, I have two .rda files that I need to include in a package. I've placed them both in a data directory after save() the are around 150Kb each. When I try to check the package I get the following warning Warning: large data file(s) saved inefficiently: size ASCII compress zagoskin.rda 137Kb FALSE none Note: significantly better compression could be obtained
2011 Dec 06
1
warning for inefficiently compressed datasets
Hi, Recently added to doc/NEWS.Rd: 'R CMD check' now gives a warning rather than a note if it finds inefficiently compressed datasets. With 'bzip2' and 'xz' compression having been available since R 2.10.0, there is no excuse for not using them. Why isn't a note enough for this? Generally speaking, warnings are for things that are dangerous, or unsafe,
2017 Sep 12
0
Load R data files
The object you load has the same name as the object you saved. In this case datahs0csv and not the name of the file sans .rda On Di., 12. Sep. 2017, 21:26 AbouEl-Makarim Aboueissa < abouelmakarim1962 at gmail.com> wrote: > Dear All: > > > It was saved, but there was a space somewhere. So it works for me now. > > I do have another similar problem. > > I saved an R
2004 Dec 23
2
Importing csv files
There is a recurring need for importing large csv files quickly. David Baird's dataload is a standalone program that will directly create .rda files from .csv (it also handles many other conversions). Unfortunately dataload is no longer publicly available because of some kind of relationship with Stat/Transfer. The idea is a good one, though. I wonder if anyone would volunteer to