Displaying 20 results from an estimated 3000 matches similar to: "load windows file in linux"
2005 Jun 30
2
ts.plot data labels
Dear R users,
I am trying to put labels on the data points on a ts.plot.
I have tried:
ts.plot(df.ts, gpars=list(xy.labels=colnames(df.ts), xlab="group", ylab="level"))
which plots fine but gives no labels
xy.labels=TRUE
also plots without labels
I also tried using text() but got an error
Error in xy.coords(x, y, recycle = TRUE) :
Argument "x" is
2005 Aug 12
1
access assigned objects
Dear list,
I am trying to write a function that will make a matrix of each row of a data frame (4 columns), calculate a fisher.test on each resulting matrix and assign a vector of the p-values. I have gotten through making the matrices, but cannot calculate the fisher.test.
fm<-function(x)
{dfrow<-nrow(x)
mm <- vector("list", dfrow)
for(i in 1:dfrow)
mm[[i]] <-
2007 Oct 10
2
slow load() in R2.6.0
I'm encountering excruciatingly slow load times for character vectors in
R 2.6.0-- up to 30sec for a 15K file that contains a no-attributes
character vector of length ~1e4 and object size ~0.5MB. In R 2.5.1,
repeated loads of the same set of files are near-instantaneous.
The problem is proving tricky to reproduce consistently from scratch, so
I have attached the 3 files used in the examples
2006 Oct 06
2
Extreme slowdown with named vectors. A bug?
Tried the following with R --vanilla on the Rv2.4.0 release (see
details at the end). I think the script and its comments speaks for
itself, but the outcome is certainly not wanted.
for (n in 58950:58970) {
cat("n=", n, "\n", sep="");
# Clean up first
rm(names, x, y); gc();
# Create a named vector of length n
# Try with format "%5d" and it
2010 Oct 05
2
[LLVMdev] gold-plugin build errors
I tried to build the gold plugin and receive the errors posted below. I
checked out the gold plugin using "cvs -z 9 -d
:pserver:anoncvs at sourceware.org:/cvs/src co src," as documented on
llvm.org. Is that version bleeding edge and perhaps not stable?
The first errors have to do with the libtool version. I have libtool
2.2.6 on my system, which is what the README-maintainer-mode
2007 Aug 11
1
LDA and RDA: different training errors
Hello
I try to fit a LDA and RDA model to the same data, which has two classes.
The problem now is that the training errors of the LDA model and the
training error of the RDA model with alpha=0 are not the same. In my
understanding this should be the case. Am I wrong? Can someone explain what
the reason for this difference could be?
Here my code:
LDA model:
===========
% x is a dataframe
tmp =
2008 Aug 22
1
save() should not overwrite a file if an error occurs (PR#12583)
If save() fails because an object is not found,
it should not overwrite an existing file.
> a <- 1:9
> save(a, file = "a.rda")
> rm(a)
> load("a.rda")
> a
[1] 1 2 3 4 5 6 7 8 9
> rm(a)
> save(a, file = "a.rda")
Error in save(a, file = "a.rda") : object 'a' not found
> load("a.rda")
Error in
2006 Mar 24
2
[PATCH] qemu pcnet emulation fixes
The attached patch to the qemu emulation of the pcnet hardware fixes
several problems. It will now only read and write a transmit or receive
descriptor once. It will correctly handle transmitting frames with more
than two fragments. It will discard oversize frames instead of
corrupting memory. I have tested all the changes I have made and even
seen an improvement in receive performance from
2007 Feb 27
2
RDA and trend surface regression
Dear all,
I'm performing RDA on plant presence/absence data, constrained by
geographical locations. I'd like to constrain the RDA by the "extended
matrix of geographical coordinates" -ie the matrix of geographical
coordinates completed by adding all terms of a cubic trend surface
regression- .
This is the command I use (package vegan):
>rda(Helling ~
2002 Feb 22
2
R gnome produces errors?: a clue!
I've got something that might be a clue to understand
the strange behaviour of lda in R --gui="gnome":
in R --vanilla:
library(MASS)
load("mod23puriflda.rda")
> a <- lda.default(x=mod23puriflda[,2:5],grouping=mod23puriflda[,6],CV=F)
works fine. I save lda.default as a local rda file:
> milda.nognome <- lda.default
>
2009 Feb 12
1
Different labels for subsets of points in a PCA or RDA biplot
I've tried a few things both with prcomp(), and rda() and its friends
in vegan (including biplot.rda and ordiplot), but can't find a
solution. I'd like to associate subsets of the points in a resulting
biplot ("sites" in the rda object) with different plotting colors/text
styles to emphasize certain sets of points. I can't figure out how to
keep the arrows (for
2012 May 16
1
Merging multiple data sets
Hello R user,
I have four data sets in dir "D:/Bharat Warule/Rdata_file" which are
output_data_prod_1.rda, output_data_prod_2.rda, output_data_prod_3.rda,
output_data_prod_4.rda.
Each data set is huge size like number of rows 343297 and columns are near
to 50.
For example:
x1 <- data.frame(x11=c(1,2,3,4,5),x112=c(10,10,10,10,10))
x2 <-
2013 Apr 24
1
RDA permutest envfit
Dear all,
I did a RDA and when I looked to the signification of the test with permutest, the output was non-significant. But when I used the envfit function, some of the vectors are significant. All the test's conditions are respected. What it means? Is it an error in the script?
Commands and output:
> permutest(rda.ind, perm=999, first=TRUE)
Permutation test for rda
Call:
2002 Jan 12
1
R-1.4.0 delay()
R-1.4.0 seems to have changed the way delay() works, or perhaps it is
the way library() works that nullifies the advantages of using delay()
in a package.
It seems that the process of loading a package evaluates each object
provided by that package, and so in particular, the delay()ed object is
evaluated and brought into memory.
For example generating a test example, where test/R/functions.s is:
2013 Mar 23
1
sysdata.rda vs. rda files in data directory
Dear developeRs,
my package FrF2.catlg128 holds large catalogues and is supposed to gain
additional ones. All the catalogues are intended for the user.
So far, the catalogues were stored in the data directory, and LazyData
was "no". I understand that this is not considered wise any more (if it
ever was), so that I want to change to LazyData "yes" with the next
release
2017 Sep 12
2
Load R data files
Dear All:
It was saved, but there was a space somewhere. So it works for me now.
I do have another similar problem.
I saved an R data file
save(datahs0csv,file="
F:\Fall_2017\5-STA574\2-Notes\1-R\1-R_new\chapter4-Entering_Data/datahs0csv2
.rda")
*The new R data file "*datahs0csv2.rda*" is in the directory.*
I tried to load the file "" to R, but I got an error
2012 Aug 16
1
Compressing data for package builds
Hi,
I have two .rda files that I need to include in a package. I've placed
them both in a data directory
after save() the are around 150Kb each.
When I try to check the package I get the following warning
Warning: large data file(s) saved inefficiently:
size ASCII compress
zagoskin.rda 137Kb FALSE none
Note: significantly better compression could be obtained
2011 Dec 06
1
warning for inefficiently compressed datasets
Hi,
Recently added to doc/NEWS.Rd:
'R CMD check' now gives a warning rather than a note if it finds
inefficiently compressed datasets. With 'bzip2' and 'xz' compression
having been available since R 2.10.0, there is no excuse for not
using them.
Why isn't a note enough for this?
Generally speaking, warnings are for things that are dangerous,
or unsafe,
2017 Sep 12
0
Load R data files
The object you load has the same name as the object you saved. In this case
datahs0csv and not the name of the file sans .rda
On Di., 12. Sep. 2017, 21:26 AbouEl-Makarim Aboueissa <
abouelmakarim1962 at gmail.com> wrote:
> Dear All:
>
>
> It was saved, but there was a space somewhere. So it works for me now.
>
> I do have another similar problem.
>
> I saved an R
2004 Dec 23
2
Importing csv files
There is a recurring need for importing large csv files quickly. David
Baird's dataload is a standalone program that will directly create .rda
files from .csv (it also handles many other conversions). Unfortunately
dataload is no longer publicly available because of some kind of
relationship with Stat/Transfer. The idea is a good one, though. I
wonder if anyone would volunteer to