similar to: help with random forest package

Displaying 20 results from an estimated 6000 matches similar to: "help with random forest package"

2009 Apr 12
3
Running random forest using different training and testing schemes
Hi, I would like to run random Forest classification algorithm and check the accuracy of the prediction according to different training and testing schemes. For example, extracting 70% of the samples for training and the rest for testing, or using 10-fold cross validation scheme. How can I do that? Is there a function? Thanks a lot, Chrysanthi. [[alternative HTML version deleted]]
2004 Jul 26
5
installing problems repeated.tgz linux
Hi, i try several possibilities adn looking in the archive, but didn't getting success to install j.lindsey's usefuel "library repeated" on my linux (suse9.0 with kernel 2.6.7,R.1.9.1) P.S. Windows, works fine Many thanks for help Christian chris at linux:/space/downs> R CMD INSTALL - l /usr/lib/R/library repeated WARNING: invalid package '-' WARNING:
2005 Sep 08
2
Re-evaluating the tree in the random forest
Dear mailinglist members, I was wondering if there was a way to re-evaluate the instances of a tree (in the forest) again after I have manually changed a splitpoint (or split variable) of a decision node. Here's an illustration: library("randomForest") forest.rf <- randomForest(formula = Species ~ ., data = iris, do.trace = TRUE, ntree = 3, mtry = 2, norm.votes = FALSE) # I am
2009 Sep 28
1
how to visualize gini coefficient in each node in RF?
Dear all, I am working with randomForest package and I am interested in examining the "Gini importance" measures that are used as a general indicator of feature relevance. Is there a possibility of getting the Gini measure that is being estimated in each tree by the output of the getTree() function? Thanks a lot, Chrysanthi [[alternative HTML version deleted]]
2010 May 17
1
Help with RandomForest
I'm working with the randomForest package and have successfully build a model. I'd like to go one step further however, and use the output from the model to construct a map using the output. My input data are spatial, and I have an independent set of rasterize maps for each of the predictor variables, to map the classification solution. Using the getTree function: >
2003 Aug 20
2
RandomForest
Hello, When I plot or look at the error rate vector for a random forest (rf$err.rate) it looks like a descending function except for a few first points of the vector with error rates values lower(sometimes much lower) than the general level of error rates for a forest with such number of trees when the error rates stop descending. Does it mean that there is a tree(s) (that is built the first in
2009 Dec 04
1
RandomForest - getTree status code
What does a status value of -3 mean when I do a regression with RF and use the getTree function? left daughter right daughter split var split point status prediction 1 2 3 11 4.721000e+03 -3 15.8489576 2 4 5 5 6.500000e+00 -3 11.3240895 3 6 7 10 6.790674e+02 -3 23.7250079 4
2002 Sep 25
5
CART vs. Random Forest
According to Dr. Breiman, the RF should be more accurate method than a single tree. However, the performance of each method seems to depend on the proprotion of outcome variable in my case. My data set is a typical classification problem (predict bad guys). When I ran both of them with different proportion of outcome variables(there's a criterion to measure the degree of bad behavior), I
2005 Mar 23
1
Gini's Importance Value Variable = Inf
Hi All, In the script below, the importance measure for column 4 (ie MeanDecreaseGini) indicated "Inf" for V7. Running the getTree command showed that "V7" had been selected at least twice in one of the trees for Random Forest. So the "Inf" command was not generated as a result of dividing the sum of the decreases by 0. Any suggestions on what may be causing the
2004 Apr 18
2
outliers using Random Forest
Hello, Does anybody know if the outscale option of randomForest yields the standarized version of the outlier measure for each case? or the results are only the raw values. Also I have notice that this measure presents very high variability. I mean if I repeat the experiment I am getting very different values for this measure and it is hard to flag the outliers. This does not happen with two other
2002 Apr 02
2
random forests for R
Hi all, There is now a package available on CRAN that provides an R interface to Leo Breiman's random forest classifier. Basically, random forest does the following: 1. Select ntree, the number of trees to grow, and mtry, a number no larger than number of variables. 2. For i = 1 to ntree: 3. Draw a bootstrap sample from the data. Call those not in the bootstrap sample the
2002 Apr 02
2
random forests for R
Hi all, There is now a package available on CRAN that provides an R interface to Leo Breiman's random forest classifier. Basically, random forest does the following: 1. Select ntree, the number of trees to grow, and mtry, a number no larger than number of variables. 2. For i = 1 to ntree: 3. Draw a bootstrap sample from the data. Call those not in the bootstrap sample the
2010 Apr 29
1
variable importance in Random Forest
HI, Dear Andy, I run the RandomFOrest in R, and get the following resutls in variable importance: What is the meaning of MeanDecreaseAccuracy and MeanDecreaseGini? I found they are raw values, they are not scaled to 1, right? Which column if most similar to the variable rel.influence in Boosting? Thanks so much! > fit$importance 0 1
2005 Oct 11
1
a problem in random forest
Hi, there: I spent some time on this but I think I really cannot figure it out, maybe I missed something here: my data looks like this: > dim(trn3) [1] 7361 209 > dim(val3) [1] 7427 209 > mg.rf2<-randomForest(x=trn3[,1:208], y=trn3[,209], data=trn3, xtest=val3[, 1:208], ytest=val3[,209], importance=T) my test data has 7427 observations but after prediction, > dim(mg.rf2$votes)
2010 Mar 01
1
Random Forest prediction questions
Hi, I need help with the randomForest prediction. i run the folowing code: > iris.rf <- randomForest(Species ~ ., data=iris, > importance=TRUE,keep.forest=TRUE, proximity=TRUE) > pr<-predict(iris.rf,iris,predict.all=T) > iris.rf$votes[53,] setosa versicolor virginica 0.0000000 0.8074866 0.1925134 > table(pr$individual[53,])/500 versicolor virginica 0.928
2012 Dec 03
2
Different results from random.Forest with test option and using predict function
Hello R Gurus, I am perplexed by the different results I obtained when I ran code like this: set.seed(100) test1<-randomForest(BinaryY~., data=Xvars, trees=51, mtry=5, seed=200) predict(test1, newdata=cbind(NewBinaryY, NewXs), type="response") and this code: set.seed(100) test2<-randomForest(BinaryY~., data=Xvars, trees=51, mtry=5, seed=200, xtest=NewXs, ytest=NewBinarY) The
2008 Oct 09
1
Dump decision trees of randomForest object
Hi, I'm using the package randomForest to generate a classifier for the exemplary iris data set: data(iris) iris.rf<-randomForest(Species~.,iris) Is it possible to print all decision trees in the generated forest? If so, can the trees be also written to disk? What I actually need is to translate the decision trees in a random forest into equivalent C++ if-then-else constructs to
2010 Feb 16
2
Random Forest
Hi, i'm using randomForest package and i have 2 questions: 1. Can i drop one tree from an RF object? 2. i have a 300 trees forest, but when i use the predict function on new data (with predict.all=TRUE) i get only 270 votes. did i do something wrong? Thanks -- View this message in context: http://n4.nabble.com/Random-Forest-tp1557464p1557464.html Sent from the R help mailing list archive at
2008 Mar 11
1
randomForest get tree
All, What purpose does the getTree function have in Random Forest? Can you graph it as you can in rpart and can it be interpreted in the same way? Helen Mills Poulos Yale School of Forestry
2002 Aug 19
4
question about Rpvm, SNOW, etc.
Dear R-devel, Inspired by Michael Li's talk at JSM, I decided to try rpvm and snow on our two linux boxes. It only took me a couple of hours of screwing around to get it working (sooner if I had RTFM). Our setup is: 2 dual PIII-866 Xeons, one with 2GB RDRAM, the other with 1.28GB RDRAM. The first machine is acting as the NIS/NFS server. both /usr and /home are exported to the second