Displaying 20 results from an estimated 5000 matches similar to: "comparing random forests and classification trees"
2004 Mar 13
4
nnet classification accuracy vs. other models
I was wandering if anybody ever tried to compare the classification
accuracy of nnet to other (rpart, tree, bagging) models. From what I
know, there is no reason to expect a significant difference in 
classification accuracy between these models, yet in my particular case
I get about 10% error rate for tree, rpart and bagging model and 80% 
error rate for nnet, applied to the same data.
Thanks.
2008 Mar 09
1
sampsize in Random Forests
Hi all,
I have a dataset where each point is assigned to a class A, B, C, or
D. Each point is also assigned to a study site. Each study site is
coded with a number ranging between 1-100. This information is stored
in the vector studySites.
I want to run randomForests using stratified sampling, so I chose the option
strata = factor(studySites)
But I am not sure how to control the number of
2013 May 16
1
3d interactive video using the rgl package
Hi all,
I've been using the 'rgl' package to visualise in 3d the water temperature recorded by a glider deployed off the coast of Australia (see snapshot attached). Using the writeWebGL function, I'm able to produce an html file of the scene with which I can then interact (e.g. zoom in/out, rotate) in my web browser.
In R, I have created another scene that includes a loop
2009 Apr 10
1
Random Forests: Question about R^2
Dear Random Forests gurus,
I have a question about R^2 provided by randomForest (for regression).
I don't succeed in finding this information.
In the help file for randomForest under "Value" it says:
rsq: (regression only) - "pseudo R-squared'': 1 - mse / Var(y).
Could someone please explain in somewhat more detail how exactly R^2
is calculated?
Is "mse"
2007 Dec 18
1
Random forests
Dear all,
I would like to use a tree regression method to analyze my dataset. I
am interested in the fact that random forests creates in-bag and
out-of-bag datasets, but I also need an estimate of support for each
split. That seems hard to do in random forests since each tree is
grown using a subset of the predictor variables.
I was thinking of setting mtry = number of predictor variables,
2012 Apr 13
1
caret package: custom summary function in trainControl doesn't work with oob?
Hi all,
I've been using a custom summary function to optimise regression model
methods using the caret package. This has worked smoothly. I've been using
the default bootstrapping resampling method. For bagging models
(specifically randomForest in this case) caret can, in theory, uses the
out-of-bag (oob) error estimate from the model instead of resampling, which
(in theory) is largely
2005 Mar 17
1
Varying grid.rect in different panels of a Lattice plot
Dear r-help,
Sleep-deprivation from having 2 youngsters under 2 around the house is 
fuzzing my brain, so please be gentle if the answer to this query is obvious!
In the example below, I'm trying to use grid.rect to add grey rectangles to 
the panels of a lattice plot to indicate which months spawning occurred of 
a (very cute) native Tasmanian fish. The fish in the two lakes spawned at 
2008 May 21
1
How to use classwt parameter option in RandomForest
Hi,
I am trying to model a dataset with the response variable Y, which has
6 levels {  Great, Greater, Greatest, Weak, Weaker, Weakest}, and
predictor variables X, with continuous and factor variables using
random forests in R. The variable Y acts like an ordinal variable, but
I recoded it as factor variable.
I ran a simulation and got OOB estimate of error rate 60%. I validated
against some
2009 Mar 11
2
Couple of Questions about Classification trees
So I have 2 sets of data - a training data set and a test data set. I've been
doing the analysis on the training data set and then using predict and
feeding the test data through that. There are 114 rows in the training data
and 117 in the test data and 1024 columns in both. It's actually the same
set of data split into two. The rows are made of 5 different numbers. They
do represent
2009 Apr 20
1
Random Forests: Predictor importance for Regression Trees
Hello!
I think I am relatively clear on how predictor importance (the first
one) is calculated by Random Forests for a Classification tree:
Importance of predictor P1 when the response variable is categorical:
1. For out-of-bag (oob) cases, randomly permute their values on
predictor P1 and then put them down the tree
2. For a given tree, subtract the number of votes for the correct
class in the
2009 Jun 09
3
rpart - the xval argument in rpart.control and in xpred.rpart
Dear R users,
I'm working with the rpart package and want to evaluate the performance of 
user defined split  functions.
I have some problems in understanding the meaning of the xval argument in 
the two functions rpart.control and xpred.rpart. In the former it is defined 
as the number of cross-validations while in the latter it is defined as the 
number of cross-validation groups. If I am
2004 Aug 19
3
Do you know if you can map a large minimum spanning tree in R?
Thanks Mike.
My data has longitude and latitude coords and I used distAB {clim.pact}
then mst {ape} to calculate my minimum spanning tree.  The nodes are
telecoms sites from all over Australia.  My goal is to determine the
minimum cost of linking them via cabling, and I'm starting by
calculating the distance "as the crow flies", but will probably
eventually need to calculate the
2005 Oct 14
1
Predicting classification error from rpart
Hi,
I think I'm missing something very obvious, but I am missing it, so I
would be very grateful for help. I'm using rpart to analyse data on
skull base morphology, essentially predicting sex from one or several
skull base measurements. The sex of the people whose skulls are being
studied is known, and lives as a factor (M,F) in the data. I want to
get back predictions of gender, and
2005 Sep 08
2
Re-evaluating the tree in the random forest
Dear mailinglist members,
I was wondering if there was a way to re-evaluate the
instances of a tree (in the forest) again after I have
manually changed a splitpoint (or split variable) of a
decision node. Here's an illustration:
library("randomForest")
forest.rf <- randomForest(formula = Species ~ ., data
= iris, do.trace = TRUE, ntree = 3, mtry = 2,
norm.votes = FALSE)
# I am
2008 Mar 06
1
Rpart and bagging - how is it done?
Hi there.
I was wondering if somebody knows how to perform a bagging procedure on a
classification tree without running the classifier with weights.
Let me first explain why I need this and then give some details of what I
have found out so far.
I am thinking about implementing the bagging procedure in Matlab.  Matlab
has a simple classification tree function (in their Statistics toolbox) but
2006 Dec 11
1
hermite and Bezier splines
Does anyone know how to do hermite or
Bezier splines in R? I can find Matlab routines, but really need to 
implement them in R. Failing that I'd be interested in other 
conformal splines. I need to smooth and interpolate animal tracking data.
Antarctic Wildlife Research Unit
School of Zoology
University of Tasmania
PO Box 252-05
Hobart
TAS, 7001
Australia
Phone:  (0)3 6226 2645
Fax:    
2004 Jan 06
2
comparing classification methods: 10-fold cv or leaving-one-out ?
Hi
what would you recommend to compare classification methods such as LDA,
classification trees (rpart), bagging, SVM, etc:
10-fold cv (as in Ripley p. 346f)
or
leaving-one-out (as e.g. implemented in LDA)?
my data-set is not that huge (roughly 200 entries)
many thanks for a hint
Christoph
-- 
Christoph Lehmann <christoph.lehmann at gmx.ch>
2004 Jun 21
2
sunrise, sunset calculation
Are there any functions available to calculate sunrise and sunset times 
for given latitude,longitude and dates?
If not, I'll appreciatte any pointers to C code I could use/port.
Thanks,
Angel
2008 Feb 25
1
To get more digits in precision of predict function of randomForests
Hi,
I am using randomForests for a classification problem. The predict
function in the randomForest library, when asked to return the
probabilities, has precision of two digits after the decimal. I need
at least four digits of precision for the predicted probabilities. How
do I achieve this?
Thank you,
Nagu
2018 Jan 22
2
Random Forests
Muchas gracias Carlos, como siempre.
Es raro que se me pasase. En su momento miré todos los argumentos del  
RF, como hago siempre, pero ese lo había olvidado. La verdad es que  
funcionaba estupendamente, pero me parecía extraño. Aunque dado que  
los RF no sobreajustan, no hay problema con que sus árboles sean todo  
lo grandes que quieras. Lo he testado con una base de datos externa y  
explica