Displaying 20 results from an estimated 100000 matches similar to: "rpart or mvpart"
2008 Feb 29
1
barplot and pca plot in mvpart/rpart
Hello,
I'm using the R package called mvpart, which is about the multivariate
regression trees.
The function I wrote is:
mrt1<- mvpart(coefmat~sChip+sScreen+sMem,data=mixdata, xv="pick",
plot.add=TRUE,uniform=TRUE,which=4,all=TRUE,xadj=2,yadj=2,rsq=TRUE,big.pts=TRUE,wgt.ave.pca=TRUE,legend=TRUE,bars=F,
pca=TRUE)
where "coefmat" is a matrix(of dimension N*K) to store
2006 Dec 28
3
CV by rpart/mvpart
Dear R-list,
I am using the rpart/mvpart-package for selecting a right-sized regression tree by 10-fold cross-validation. My question: Is there a possibility to find out for every observation in which of the ten folds it is lying? I want to use the same folds for validating another regression method (moving averages) in order to choose the better one.
Thanks a lot,
Pedro
2008 Oct 01
0
xpred.rpart() in library(mvpart)
R-users
E-mail: r-help@r-project.org
Hi! R-users.
http://finzi.psych.upenn.edu/R/library/mvpart/html/xpred.rpart.html
says:
data(car.test.frame)
fit <- rpart(Mileage ~ Weight, car.test.frame)
xmat <- xpred.rpart(fit)
xerr <- (xmat - car.test.frame$Mileage)^2
apply(xerr, 2, sum) # cross-validated error estimate
# approx same result as rel. error from printcp(fit)
apply(xerr, 2,
2010 Aug 13
3
Delete rpart/mvpart cross-validation output
Dear all,
I was wondering if there is a simple way to avoid printing the multiple
cross-validation automatic output to the console of recursive partitionning
functions like rpart or mvpart. For example...
> data(spider)
>
mvpart(data.matrix(spider[,1:12])~herbs+reft+moss+sand+twigs+water,spider,xv="1se",xvmult=100)
*X-Val rep : 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
2004 Feb 17
0
New package -- mvpart
The package mvpart is now available.
mvpart includes partitioning based on (1) multivariate numeric responses and
(2) dissimilarity matrices.
The package mvpart is a modification of rpart --
-- authors of original: Terry M Therneau and Beth Atkinson
<atkinson at mayo.edu>, and
R port of rpart Brian Ripley <ripley at stats.ox.ac.uk>.
Includes some modified routines from vegan --
2004 Feb 17
0
New package -- mvpart
The package mvpart is now available.
mvpart includes partitioning based on (1) multivariate numeric responses and
(2) dissimilarity matrices.
The package mvpart is a modification of rpart --
-- authors of original: Terry M Therneau and Beth Atkinson
<atkinson at mayo.edu>, and
R port of rpart Brian Ripley <ripley at stats.ox.ac.uk>.
Includes some modified routines from vegan --
2010 Feb 26
2
Error in mvpart example
Dear all,
I'm getting an error in one of the stock examples in the 'mvpart' package. I tried:
require(mvpart)
data(spider)
fit3 <- rpart(gdist(spider[,1:12],meth="bray",full=TRUE,sq=TRUE)~water+twigs+reft+herbs+moss+sand,spider,method="dist") #directly from ?rpart
summary(fit3)
...which returned the following:
Error in apply(formatg(yval, digits - 3), 1,
2009 Mar 23
1
mvpart error
Hello all,
When attempting a classification tree using mvpart, I get the following
error:
> thesis2.mvp=mvpart(bat_sp~., data=alltrees.df)
Error in all(keep) :
unused argument(s) (c(TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE,
TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE,
TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE, TRUE,
TRUE, TRUE, TRUE,
2010 Mar 12
1
using xval in mvpart to specify cross validation groups
Dear R's
I'm trying to use specific rather than random cross-validation groups
in mvpart.
The man page says:
xval Number of cross-validations or vector defining cross-validation groups.
And I found this reply to the list by Terry Therneau from 2006
The rpart function allows one to give the cross-validation groups explicitly.
So if the number of observations was 10, you could use
2008 Feb 29
1
controlling for number of elements in each node of the tree in mvpart
Still about the mvpart.
Is there any way I can control for the number of elements in each node
in the function mvpart? Specifically, how can I ask partition to
ignore node with elements less than 10?
Thanks!
-Shu
2005 Aug 08
2
INDVAL and mvpart
Hi,
I'd like to perform Dufrene-Legendre Indicator Species Analysis for
a multivariate regression tree. However I have problems with arguments
of duleg(veg,class,numitr=1000)function. How to obtain a vector of
numeric class memberships for samples, or a classification object
returned from mvpart?
thanks in advance
--
Best regards,
Agnieszka Strzelczak
2008 Sep 16
1
1-SE rule in mvpart
Hello,
I'm using mvpart option xv="1se" to compute a regression tree of good size
with the 1-SE rule.
To better understand 1-SE rule, I took a look on its coding in mvpart, which
is :
Let z be a rpart object ,
xerror <- z$cptable[, 4]
xstd <- z$cptable[, 5]
splt <- min(seq(along = xerror)[xerror <= min(xerror) + xvse * xstd])
I interprete this as following: the
2009 Mar 15
0
mvpart error - is.leaf
Hello,
When trying to run mvpart either specifying my own parameters or using the
defaults, I get the following error:
Error in all(is.leaf) :
unused argument(s) (c(FALSE, TRUE, FALSE, FALSE, TRUE, TRUE, TRUE))
As far as I can tell, is.leaf is part of the dendrogam package, so I'm
assuming there's some problem with the graphical parameters. However running
same formula and data
2012 Apr 23
1
change color scheme in mvpart
Hello everyone, I am currently using the mvpart package and would like to change the color scheme it uses, and was hoping someone could help me out. All of the papers I have found have used a grayscale but I can't seem to figure out how they did that! Currently, mvpart plots barplots in a repeating sequence of 3 shades of blue. So if you have 6 response variables the same shade of blue is used
2011 Sep 13
1
mvpart analyses with covariables
Hi all,
I am fairly new to R and I am trying to run mvpart and create a MRT using
explanatory variables and covariables. I've been following the procedures in
Numerical Ecoogy with R.
The command (no covariables) which works fine -
ABUNDTMRT <- mvpart(abundance ~
.,factors,margin=0.08,cp=0,xv="1se",xval=nrow(abundance),xvmult=100,which=4)
where abundance is 4th root
2006 Oct 17
1
Some questions on Rpart algorithm
Hello:
I am using rpart and would like more background on how the splits are made
and how to interpret results - also how to properly use text(.rpart). I have
looked through Venables and Ripley and through the rpart help and still have
some questions. If there is a source (say, Breiman et al) on decision trees
that would clear this all up, please let me know. The questions below
pertain to a
2014 Aug 13
1
Request to review a patch for rpart
Dear list
For my work, it would be helpful if rpart worked seamlessly with an
empty model:
library(rpart); rpart(formula=y~0, data=data.frame(y=factor(1:10)))
Currently, an unrelated error (originating from na.rpart) is thrown.
At some point in the near future, I'd like to release a package to CRAN
which uses rpart and relies on that functionality. I have prepared a
patch (minor
2009 Dec 14
1
RPART - printing full splitting rule number on tree plot
Dear R-users
I am using RPART package to get regression trees. However having trouble getting the text function to put the full splitting rule number on the plot, instead to puts it in scientific notation. When a covariate has 1e4 or greater number of digits then the splitting rule number displayed on the plot is in scientific notation. But print.rpart displays the splitting rules in full.
2008 Jul 03
1
cross-validation in rpart
Hello list,
I'm having a problem with custom functions in rpart, and before I tear my
hair out trying to fix it, I want to make sure it's actually a problem. It
seems that, when you write custom functions for rpart (init, split and eval)
then rpart no longer cross-validates the resulting tree to return errors. A
simple test is to use the usersplits.R function to get a simple, custom
2012 Apr 24
0
mvpart versus SPSS
I have a question relating to mvpart, which I hope you can answer.
We recently conducted a study using TBR. In our first study, we used
"regular" TBR in SPSS to model 1 dependent variable. Note we have a
relatively small data-set of 100 cases. In SPSS, we used a minimum change of
improvement smaller than 0.000001 as a stopping rule. Also, we chose the 1SE
"rule", set the