similar to: R CMD check for packages in a bundle

Displaying 20 results from an estimated 5000 matches similar to: "R CMD check for packages in a bundle"

2009 Jun 12
1
bundle deprecation
Hi I read that bundles are to be deprecated in 2.10. The BACCO bundle contains three packages (emulator, calibrator, approximator) which I am happy to unbundle. But the 'BACCO' moniker has some considerable cachet for me in terms of recognizability (eg with grant-giving bodies), as it has become an umbrella term for a whole bunch of related statistical functionality of which the three
2005 Oct 24
0
new package bundle on CRAN: BACCO
Dear List please find on CRAN a new R bundle, BACCO, for Bayesian analysis of random functions, comprising two packages: emulator and calibrator. Package calibrator implements: ?Bayesian calibration of computer models?, M. C. Kennedy and A. O'Hagan 2001. Journal of the Royal Statistical Society B, 63(3) pp425-464 and package emulator implements: J. Oakley 2004. ?Estimating percentiles
2004 Jul 05
1
"make" error for R-1.9.1
Hello everybody. I am trying to upgrade from R-1.9.0 to R-1.9.1 on a RedHat linux 2.4.18 system. I get the following error after "tar -xvzf R-1.9.1.tgz ; cd ./R-1.9.1/ ; ./configure" and "make" : [make works for 10 minutes ... snip ...] varExp text html latex varFixed text html latex varFunc
2004 Nov 29
1
data() in data/*.R files
Hi I'm having difficulty making a package pass R CMD check. I need to read in a dataset from another package, modify it, and have the modified object available in the first package. help(require) says: The source code for a package that requires one or more other packages should have a call to 'require', preferably near the beginning of the source, and of course
2004 Sep 14
1
R-2.0.0 CMD check . and datasets
Hello everyone I'm having a little difficulty with R-2.0.0 CMD check. My field is Bayesian calibration of computer models. The problem is that I have a large collection of toy datasets, that in R-1.9.1 were specified with lines like this: x.toy <- 1:6 y.toy <- computer.model(x.toy) z.toy <- reality(x.toy) in file ./data/toys.R ; functions computer.model() and reality() are
2004 Sep 14
1
R-2.0.0 CMD check . and datasets
Hello everyone I'm having a little difficulty with R-2.0.0 CMD check. My field is Bayesian calibration of computer models. The problem is that I have a large collection of toy datasets, that in R-1.9.1 were specified with lines like this: x.toy <- 1:6 y.toy <- computer.model(x.toy) z.toy <- reality(x.toy) in file ./data/toys.R ; functions computer.model() and reality() are
2004 Sep 13
1
Rd files with "%" (was: permuting dimensions)
Professor Ripley thanks for this. Very much appreciated. The original subject line reflected my late-night conviction that the answer might involve passing a strange list to do.call(). Anyway, package magic is broken (only in R-devel, I might add) because I have a function called "%eq%". R-2.0.0 CMD check is stopping (I think) because it interprets the "%" as a
2004 Sep 08
3
do.call("[", ...) question
Hi again everyone I have an arbitrarily dimensional array "a" and a list "jj" of length length(dim(a)). The elements of jj are vectors of indexes. How do I use do.call() to extract a[ jj[[1]], jj[[2]], jj[[3]], ...] ? Toy example follows: a <- matrix(1:30,5,6) jj <- list(5:1,6:1) I want the following a[ jj[[1]],jj[[2]] ] How do I do this? OBAttempts:
2004 Jun 11
4
rownames of single row matrices
Hi I want to extract rows of a matrix, and preserve rownames even if only one row is selected. Toy example: R> a <- matrix(1:9,3,3) R> rownames(a) <- letters[1:3] R> colnames(a) <- LETTERS[1:3] R> a A B C a 1 4 7 b 2 5 8 c 3 6 9 Extract the first two rows: R> wanted <- 1:2 R> a[wanted,] A B C a 1 4 7 b 2 5 8 rownames come through fine. Now extract just
2004 Nov 18
0
Fwd: Re: 3d scatter plot with drop line
Hi try this: p3dpairs <- function(x,x1, xlim=NULL,ylim=NULL,zlim=NULL,col=par("col"), pch=par("pch"), cex=par("cex"), ...){ if(is.matrix(x)){ z <- x[,3] y <- x[,2] x <- x[,1] } if(is.matrix(x1)){ z1 <- x1[,3] y1 <- x1[,2] x1 <- x1[,1] } if(missing(zlim)) { z.grid <-
2003 Oct 24
2
x[c(1,2,1)] <- 1:3
Hi everyone. look at this: x <- 1:4 x[c(1,2,1)] <- (1:3) print(x[1]) I get 3, but isn't NA more appropriate? [1 would be as sensible]. FWIW, the equivalent Fortran 95 statement is flagged as an error. R-intro, section 2.7, says that in such cases the assignment is carried out "in order" which might support getting 3. To my way of thinking, the concept of "in
2011 Jan 12
1
GLMM with lme4 and octopus behaviour
Hi all, First time poster and a relatively new R user, I'm beginning analysis for my masters degree. I'm doing a bit of work on octopus behaviour, and while it's been fascinating, the stats behind it are a bit beyond my grasp at the moment. I was hoping that somebody with more experience my be able to look at my example and offer their wisdom, much to my appreciation :-) At the most
2004 May 27
2
block diagonal matrix function
Hello List I have just written a little function that takes two matrices as arguments and returns a large matrix that is composed of the two input matrices in upper-left position and lower-right position with a padding value everywhere else. (function definition and toy example below). I need nonsquare matrices and rowname() and colname() inherited appropriately. Two questions: (1) Is there a
2004 Mar 12
1
another do.call() problem.
Hi everyone suppose I have a <- array(1:256,rep(4,4)) and want to access a[1,2,3,1] by the vector c(1,2,3,1). As per yesterday, I can use do.call(): a[1,2,3,1] == do.call("[",c(list(a),c(1,2,3,1))) Now how do I apply the above technique (or indeed any other technique!) to get a[1,2,3,] [1] 37 101 165 229 from a vector like c(1,2,3,0) or c(1,2,3,NULL) or c(1,2,3,NA)?
2004 May 25
2
vectorize an expression
Hi guys. Another thing I cannot vectorize: I have an array "a", of size 5-by-2, for example, of 5 2D vectors. I also have a distance function that computes the distance between two vectors (usual Euclidean distance is a good example but I have other metrics I want to use as well). I want a 5-by-5 array with the [i,j]th element being the distance from a[i,] to a[j,] To Wit: a
2004 Mar 10
3
aperm() and as.list() args to "["
Hi everyone. I'm playing with aperm(): a <- 1:24 dim(a) <- c(2,3,2,2) permutation <- c(1,2,4,3) b <- aperm(a,permutation) So if my understanding is right, a[1,3,2,1] == b[c(1,3,2,1)[permutation] ] but this isn't what I want because the RHS evaluates to a vector, and I am trying to identify a single element of b. How do I modify the RHS to give what I want? Following
2004 Nov 15
2
argument rationalization
Hi I am writing a bunch of functions that take two, three or four arguments. These functions operate on vectors of the same length; but I want the function to behave sensibly if one or more arguments are scalars. "+" does this for two arguments: "+"(1:10,3) # interpreted as "+"(1:10,rep(3,10)) But my functions can take more arguments. Say f() takes three:
2004 Oct 06
3
crossprod vs %*% timing
Hi the manpage says that crossprod(x,y) is formally equivalent to, but faster than, the call 't(x) %*% y'. I have a vector 'a' and a matrix 'A', and need to evaluate 't(a) %*% A %*% a' many many times, and performance is becoming crucial. With f1 <- function(a,X){ ignore <- t(a) %*% X %*% a } f2 <- function(a,X){ ignore <-
2004 Nov 17
3
3d scatter plot with drop line
This is a follow up to my question from yesterday. I want to do in R what is called a "3d scatter plot with drop lines" in S-PLUS. Basically, it's a 3dscatterplot with lines connecting the x-y grid to the z points. The lines give a better perspective on the shape of the data surface. How to? Joel Bremson UC Davis Statistics
2002 Jan 14
1
new R documentation on CRAN
Dear R community A few weeks ago, I uploaded a small text file called R-and-octave.txt to the contributed docs section of CRAN. This file details octave/matlab commands and their (near) equivalents in R (Matlab is a widely-used high-level graphics/mathematics tool and octave a free clone). Someone has just pointed out to me that I never announced its existence to anyone, hence this email (I