Displaying 20 results from an estimated 8000 matches similar to: "Question on implementing Random Forests scoring"
2009 Apr 10
1
Random Forests: Question about R^2
Dear Random Forests gurus,
I have a question about R^2 provided by randomForest (for regression).
I don't succeed in finding this information.
In the help file for randomForest under "Value" it says:
rsq: (regression only) - "pseudo R-squared'': 1 - mse / Var(y).
Could someone please explain in somewhat more detail how exactly R^2
is calculated?
Is "mse"
2008 Mar 09
1
sampsize in Random Forests
Hi all,
I have a dataset where each point is assigned to a class A, B, C, or
D. Each point is also assigned to a study site. Each study site is
coded with a number ranging between 1-100. This information is stored
in the vector studySites.
I want to run randomForests using stratified sampling, so I chose the option
strata = factor(studySites)
But I am not sure how to control the number of
2017 Jul 07
1
Scoring and Ranking Methods
Hi,
I am doing predictive modelling of Multivariate Time series Data of a Motor
in R using various models such as Arima, H2O.Randomforest, glmnet, lm and
few other models.
I created a function to select a model of our choice and do prediction.
Model1 <- function(){
..
return()
}
Model2 <- function(){
...
return()
}
Model3 <- function(){
...
return()
}
main <-
2003 Jul 09
2
CFP: CART Data Mining Conference 2004
Apologies for cross posting....
---------------------------------------------------------------------
CART Data Mining'04: First International CART(R) Conferences
Focusing on the Data Mining technology of
Leo Breiman, Jerome Friedman, Richard Olshen, Charles Stone
(CART, MARS(R), TreeNet(tm), PRIM(tm)...)
First Call For submissions
2012 Dec 03
2
Different results from random.Forest with test option and using predict function
Hello R Gurus,
I am perplexed by the different results I obtained when I ran code like
this:
set.seed(100)
test1<-randomForest(BinaryY~., data=Xvars, trees=51, mtry=5, seed=200)
predict(test1, newdata=cbind(NewBinaryY, NewXs), type="response")
and this code:
set.seed(100)
test2<-randomForest(BinaryY~., data=Xvars, trees=51, mtry=5, seed=200,
xtest=NewXs, ytest=NewBinarY)
The
2007 Sep 17
1
random Forests
Hi,
I am new to R and have a specific question about the randomForest package and the saving of trees and scoring.
1) I am looking to save the trees and score at a later time. Is there a way to load the saved trees and use the predict function? Can objects be saved and loaded i.e. the randomForest function call? I dont want to have to rerun trees. Hopefully this applies to any stat type
2009 Apr 13
2
Random Forests Variable Importance Question
I am trying to use the random forests package for classification in R.
The Variable Importance Measures listed are:
-mean raw importance score of variable x for class 0
-mean raw importance score of variable x for class 1
-MeanDecreaseAccuracy
-MeanDecreaseGini
Now I know what these "mean" as in I know their definitions. What I
want to know is how to use them.
What I am trying to
2009 Apr 20
1
Random Forests: Predictor importance for Regression Trees
Hello!
I think I am relatively clear on how predictor importance (the first
one) is calculated by Random Forests for a Classification tree:
Importance of predictor P1 when the response variable is categorical:
1. For out-of-bag (oob) cases, randomly permute their values on
predictor P1 and then put them down the tree
2. For a given tree, subtract the number of votes for the correct
class in the
2007 Dec 18
1
Random forests
Dear all,
I would like to use a tree regression method to analyze my dataset. I
am interested in the fact that random forests creates in-bag and
out-of-bag datasets, but I also need an estimate of support for each
split. That seems hard to do in random forests since each tree is
grown using a subset of the predictor variables.
I was thinking of setting mtry = number of predictor variables,
2010 Oct 22
2
Random Forest AUC
Guys,
I used Random Forest with a couple of data sets I had to predict for binary
response. In all the cases, the AUC of the training set is coming to be 1.
Is this always the case with random forests? Can someone please clarify
this?
I have given a simple example, first using logistic regression and then
using random forests to explain the problem. AUC of the random forest is
coming out to be
2012 Jan 27
1
Bivariate Partial Dependence Plots in Random Forests
Hello,
I was wondering if anyone knew of an R function/R code to plot bivariate
(3 dimensional) partial dependence plots in random forests (randomForest
package).
It is apparently possible using the rgl package
(http://esapubs.org/archive/ecol/E088/173/appendix-C.htm) or there may
be a more direct function such as the pairplot() in MART (multiple
additive regression trees)?
Many
2008 Oct 21
3
Exclude rows in table
Dear R-Help,
I have the following type of table (with number of rows = 4765) and want
to exclude each row where Net=0. Could this be done in a simple way?
Thanks in advance
Jim
> A
Business.Unit Event1 Net Date
1 General Fraud 170.000 2006-01-01
2 General Fraud 100.000 2007-11-09
3 General Fraud 486.286
2005 May 09
1
Random Forests 4.5-10 varImpPlot (PR#7844)
Full_Name: Daniel Normolle
Version: 2.0.1
OS: Linux/Fedora Core 3
Submission from: (NULL) (141.214.17.5)
varImpPlot in Random Forests 4.5-10 produces the error "incorrect number of
subscripts on matrix" (and no plot) when applied to a randomForest object. This
error did not occur with 4.5-4 or earlier versions.
2008 Jun 02
1
Random Forests regression by strata
Hello,
I'm trying to sample in Random Forests by a factor, but it is a regression problem and I can't figure out how to do this (I can only see how to sample by strata in classification).
Thanks
Jesse Lasky
2006 Jul 23
1
Iterated Data Input/Output with Random Forests
Hi,
I am currently writing code to input a few thousand files, run them through the
Random Forests package, and then output corresponding results.
When I use the code below:
zz<-textConnection("ex.lm.out", "w")
sink(zz)
2012 May 11
2
Random forests prediction
Hi all,
I have a strange problem when applying RF in R.
I have a set of variables with which I obtain an AUC of 0.67.
I do have a second set of variables that have an AUC of 0.57.
When I merge the first and second set of variables, the AUC becomes 0.64.
I would expect the prediction to become better as I add variables that do
have some predictive power?
This is even more strange as the AUC
2007 Aug 24
2
Variable Importance - Random Forest
Hello,
I am trying to explore the use of random forests for classification and
am certain about the interpretation of the importance measurements.
When having the option "importance = T" in the randomForest call, the
resulting 'importance' element matrix has four columns with the
following headings:
0 - mean raw importance score of variable x for class 0 (where
2007 Jan 29
3
comparing random forests and classification trees
Hi,
I have done an analysis using 'rpart' to construct a Classification Tree. I
am wanting to retain the output in tree form so that it is easily
interpretable. However, I am wanting to compare the 'accuracy' of the tree
to a Random Forest to estimate how much predictive ability is lost by using
one simple tree. My understanding is that the error automatically displayed
by the two
2013 Oct 18
3
fraud detection
hello everyone. i am concerned about security to the PBX and i would like
to discuss different fraud detection methods.
Apart from making everything to secure the PBX (latest patches, iptables,
firewalls, no outside users, strongs passwds,...) i would like to find out
if there are any fraud detection techniques.
As for my setup i do have a PBX running asterisk 11.4 and it has 3 sip
trunks (over
2004 Mar 02
1
some question regarding random forest
Hi,
I had two questions regarding random forests for regression.
1) I have read the original paper by Breiman as well as a paper
dicussing an application of random forests and it appears that the one
of the nice features of this technique is good predictive ability.
However I have some data with which I have generated a linear model
using lm(). I can get an RMS error of 0.43 and an R^2 of