similar to: Simplifying Sweave graphics

Displaying 20 results from an estimated 4000 matches similar to: "Simplifying Sweave graphics"

2008 Feb 25
3
How to include the documentation of a function in a Sweave document?
Dear R-help, I would like to include the documentation of an R function in an *.rnw document processed by Sweave. Because I'm sharing my *.rnw files with colleagues under Linux and Windows (I'm on Mac OS X), I would like a pure R solution. The naive approach doesn't work, because Sweaving this *.rnw file: -------- tmp.rnw -------- \documentclass{article} \begin{document}
2007 Nov 02
1
Sweave, and long lasting scripts
Dear R user, I am using Sweave to write my reports with R code. Unlucly my code take a lot of time to run, and I don't want to run it multiple time, if possible, as I am still developping the tools. It seems that a sweave run does not save the enviroment at the end of the process, is there a way to avoid to lunch multiple time the same commands? maybe it could be possible to store chunck of
2008 Apr 01
1
superimpose histogram on biplot
Hi all, I've been trying to figure out how to superimpose a histogram on a biplot that shows the relative contribution of each axis. I have been using the NIPALS function (http://biomserv.univ-lyon1.fr/~dray/ files/softwares/nipals.R) to run principal component analyses. Here is a toy example. source("http://biomserv.univ-lyon1.fr/~dray/files/softwares/nipals.R")
2007 Nov 23
4
PCA with NA
Dear all, (Mac OS X 10.4.11, R 2.6.0) I have a quantitative dataset with a lot of Na?s in it. So many, that it is not possible to delete all rows with NA?s and also not possible, to delete all variables with NA?s. Is there a function for a principal component analysis, that can deal with so many NA?s. Thanks in advance Birgit Birgit Lemcke Institut f?r Systematische Botanik
2007 Nov 21
3
shrink a dataframe for plotting
I get tables with millions of rows. For plotting to a screen-size jpg, obviously just about 1000 points are enough. Instead of feeding plot() the original millions of rows, I'd rather shrink the original dataframe, using some kind of the following interpolation: -- split dataframe into chunks of N rows each, e.g. 1000 rows each -- compute average for each column -- issue one new row
2004 Oct 22
4
Evaluate a function for various value of parameters
Hello list, I have a problem ... and do not know how to solve it. I would like create a function that estimate the quality of the fit of different functions with different values of parameters. This problem is related to the following one: I would like to create a function "evaluatemyfunction" which evaluate a function f for different values of parameters. For instance, if
2005 Nov 11
1
A 'sweave' strange problem !!!
Hello list, I have found a problem (bug?) with Sweave. I hope that someone could help me. Try this little example : \documentclass[a4paper]{article} \title{toto} \author{toto} \begin{document} \maketitle <<ni1, fig=T, eval=TRUE, echo=TRUE, debug=TRUE, results=verbatim, include =FALSE, width=7, height=7>>= a<-rnorm(1)+1 a plot(1:10) @ patati <<ni2, fig=F, eval=TRUE,
2007 Nov 16
1
Efficient way to compute power of a sparse matrix
Dear all, I would like to compute power of a square non symmetric matrix. This is a part of a simulation study. Matrices are quite large (e.g., 900 by 900), and contains many 0 (more than 99 %). I have try the function mtx.exp of the Biodem package: library(Biodem) m <- matrix(0, 900, 900) i <- sample(1:900, 3000, replace = T) j <- sample(1:900, 3000, replace = T) for(x in 1:3000)
2001 Nov 06
2
Canonical Correspondence analysis-CoCoAn package
Hi R-users, I am new to R environment.I want to carry out a correspondence analysis on a contigency table with 64 columns and 298 observation of Environmental data.In many cells of the contigency table the frequency is just the value '1'. I got the following error. > CAIV(Eplankton) Error in if (L[i, j] < 0) return("Table L must contain non-negative numbers") :
2001 May 29
3
geary statistics
Hello, i' m looking for a function to compute geary statistic for spatial correlation. Thanks. -- St?phane DRAY --------------------------------------------------------------- Biom?trie et Biologie ?volutive - Equipe "?cologie Statistique" Universite Lyon 1 - Bat 711 - 69622 Villeurbanne CEDEX - France Tel : 04 72 43 27 56 Fax : 04 78 89 27 19 04 72 43 27 57 E-mail
2002 Jul 18
2
RODBC and Excel Files
Hello, I am trying to play with RODBC library and Excel Files. In my file (doubs.xls) there are 2 spreadsheets: > library(RODBC) > connection<-odbcConnect("Excel Files") > sqlTables(connection) TABLE_CAT TABLE_SCHEM TABLE_NAME TABLE_TYPE REMARKS 1 F:\\Th?se\\R\\Doubs NA Faune$ SYSTEM TABLE NA 2 F:\\Th?se\\R\\Doubs NA Milieu$
2006 Jul 01
3
Installation, permissions of /usr/local/lib/R (PR#9054)
Full_Name: Jori M?ntysalo Version: 2.3.1 OS: Linux, GNU/Debian stable Submission from: (NULL) (81.197.171.182) I said ./configure --with-readline=no --with-x=no make make install and everything works except that /usr/local/lib/R/etc/ldpaths was not readable as normal user.
2007 Mar 28
2
Standardization Range
? stato filtrato un testo allegato il cui set di caratteri non era indicato... Nome: non disponibile Url: https://stat.ethz.ch/pipermail/r-help/attachments/20070328/ee4422a2/attachment.pl
2006 Dec 05
1
Cummulative Variance in Correspondence Analysis (ADE4)
Hi all: How can I calculate the cumulative variance (or variance for each component) in correspondence analysis? If were possible in ADE4 package Thank you -- Antonio Punzón Merino O__---- Instituto Español de Oceanografía c/ /'_ --- Centro Oceanográfico de Santander (*) \(*) -- Promontorio de San Martín S/N ~~~~~~~~~~ 39004-Santander; Spain PO BOX: 240 Tlf: +34 942 29 10 60 Fax: +34
2008 Apr 22
4
Modeling presence only data in R
Dear All, I have a set of environental maps and presence-only points for some species. How can I generate distributions models on R using these presence-only data? What packages and functions can I use to do so? Kind regards, Miltinho Brazil [[alternative HTML version deleted]]
2007 Nov 21
1
problem with nb2listw
Hi, I am a new student to R and I was hoping someone could help me with the error message I keep getting when I try to use the nb2listw() function. I have 45 sites and I want to get an autocorrelation value for sites within 1000m of each other. coords<-cbind(EAST,NORTH) coords<-as.matrix(coords) nb1000<-dnearneigh(coords,0,1000,longlat=TRUE) then when I try to run the
2007 Oct 22
1
grouping rows into bin ranges
Hi everyone, I have a matrix that contains 1000 replicates of a sample of a list of values. I want to group each row (row=replicate) into my defined bin ranges and then calculate the mean and stdev for each of the bin ranges (so I will have 1000 rows but ncol=number of bin ranges). I don't know how to group rows in a matrix matrix according to another vector (which is what I think I need to
2008 Feb 23
1
Standard method for S4 object
Hi the list, I am defining a new class MyClass. Shortly, I will submit a package with it. Before, I would like to know if there is a kind of "non official list" of what method a new S4 object have. More precisely, personnaly, I use 'print', 'summary' and 'plot' a lot. So for my new class, I define these 3 methods. Is there some other method that a R user can
2009 Apr 03
1
Weighted principal components analysis?
Hello R-ers, I'm trying to do a weighted principal components analysis. I couldn't find any such option with princomp or prcomp. Does anyone know of a package or way to do this? More specifically, the observations I'm working with are averages from populations of varying sizes. I thus need to weight the observations by sample size. Ideally I could apply these weights at the cell
2008 Feb 26
3
R package to perform Horn's parallel analysis
I am seeking information on whether anyone has written code to perform Horn's parallel analysis (a procedure that informs the selection of the proper number of components in PCA) in R. Thank you in advance for any help you can provide. Please respond off-list at the email address below. Karen Douglas ******************************************************************* Karen Douglas,