similar to: Question on implementing Random Forests scoring

Displaying 20 results from an estimated 8000 matches similar to: "Question on implementing Random Forests scoring"

2009 Apr 10
1
Random Forests: Question about R^2
Dear Random Forests gurus, I have a question about R^2 provided by randomForest (for regression). I don't succeed in finding this information. In the help file for randomForest under "Value" it says: rsq: (regression only) - "pseudo R-squared'': 1 - mse / Var(y). Could someone please explain in somewhat more detail how exactly R^2 is calculated? Is "mse"
2008 Mar 09
1
sampsize in Random Forests
Hi all, I have a dataset where each point is assigned to a class A, B, C, or D. Each point is also assigned to a study site. Each study site is coded with a number ranging between 1-100. This information is stored in the vector studySites. I want to run randomForests using stratified sampling, so I chose the option strata = factor(studySites) But I am not sure how to control the number of
2017 Jul 07
1
Scoring and Ranking Methods
Hi, I am doing predictive modelling of Multivariate Time series Data of a Motor in R using various models such as Arima, H2O.Randomforest, glmnet, lm and few other models. I created a function to select a model of our choice and do prediction. Model1 <- function(){ .. return() } Model2 <- function(){ ... return() } Model3 <- function(){ ... return() } main <-
2003 Jul 09
2
CFP: CART Data Mining Conference 2004
Apologies for cross posting.... --------------------------------------------------------------------- CART Data Mining'04: First International CART(R) Conferences Focusing on the Data Mining technology of Leo Breiman, Jerome Friedman, Richard Olshen, Charles Stone (CART, MARS(R), TreeNet(tm), PRIM(tm)...) First Call For submissions
2012 Dec 03
2
Different results from random.Forest with test option and using predict function
Hello R Gurus, I am perplexed by the different results I obtained when I ran code like this: set.seed(100) test1<-randomForest(BinaryY~., data=Xvars, trees=51, mtry=5, seed=200) predict(test1, newdata=cbind(NewBinaryY, NewXs), type="response") and this code: set.seed(100) test2<-randomForest(BinaryY~., data=Xvars, trees=51, mtry=5, seed=200, xtest=NewXs, ytest=NewBinarY) The
2007 Sep 17
1
random Forests
Hi, I am new to R and have a specific question about the randomForest package and the saving of trees and scoring. 1) I am looking to save the trees and score at a later time. Is there a way to load the saved trees and use the predict function? Can objects be saved and loaded i.e. the randomForest function call? I dont want to have to rerun trees. Hopefully this applies to any stat type
2009 Apr 13
2
Random Forests Variable Importance Question
I am trying to use the random forests package for classification in R. The Variable Importance Measures listed are: -mean raw importance score of variable x for class 0 -mean raw importance score of variable x for class 1 -MeanDecreaseAccuracy -MeanDecreaseGini Now I know what these "mean" as in I know their definitions. What I want to know is how to use them. What I am trying to
2009 Apr 20
1
Random Forests: Predictor importance for Regression Trees
Hello! I think I am relatively clear on how predictor importance (the first one) is calculated by Random Forests for a Classification tree: Importance of predictor P1 when the response variable is categorical: 1. For out-of-bag (oob) cases, randomly permute their values on predictor P1 and then put them down the tree 2. For a given tree, subtract the number of votes for the correct class in the
2007 Dec 18
1
Random forests
Dear all, I would like to use a tree regression method to analyze my dataset. I am interested in the fact that random forests creates in-bag and out-of-bag datasets, but I also need an estimate of support for each split. That seems hard to do in random forests since each tree is grown using a subset of the predictor variables. I was thinking of setting mtry = number of predictor variables,
2010 Oct 22
2
Random Forest AUC
Guys, I used Random Forest with a couple of data sets I had to predict for binary response. In all the cases, the AUC of the training set is coming to be 1. Is this always the case with random forests? Can someone please clarify this? I have given a simple example, first using logistic regression and then using random forests to explain the problem. AUC of the random forest is coming out to be
2012 Jan 27
1
Bivariate Partial Dependence Plots in Random Forests
Hello, I was wondering if anyone knew of an R function/R code to plot bivariate (3 dimensional) partial dependence plots in random forests (randomForest package). It is apparently possible using the rgl package (http://esapubs.org/archive/ecol/E088/173/appendix-C.htm) or there may be a more direct function such as the pairplot() in MART (multiple additive regression trees)? Many
2008 Oct 21
3
Exclude rows in table
Dear R-Help, I have the following type of table (with number of rows = 4765) and want to exclude each row where Net=0. Could this be done in a simple way? Thanks in advance Jim > A Business.Unit Event1 Net Date 1 General Fraud 170.000 2006-01-01 2 General Fraud 100.000 2007-11-09 3 General Fraud 486.286
2005 May 09
1
Random Forests 4.5-10 varImpPlot (PR#7844)
Full_Name: Daniel Normolle Version: 2.0.1 OS: Linux/Fedora Core 3 Submission from: (NULL) (141.214.17.5) varImpPlot in Random Forests 4.5-10 produces the error "incorrect number of subscripts on matrix" (and no plot) when applied to a randomForest object. This error did not occur with 4.5-4 or earlier versions.
2008 Jun 02
1
Random Forests regression by strata
Hello, I'm trying to sample in Random Forests by a factor, but it is a regression problem and I can't figure out how to do this (I can only see how to sample by strata in classification). Thanks Jesse Lasky
2006 Jul 23
1
Iterated Data Input/Output with Random Forests
Hi, I am currently writing code to input a few thousand files, run them through the Random Forests package, and then output corresponding results. When I use the code below: zz<-textConnection("ex.lm.out", "w") sink(zz)
2012 May 11
2
Random forests prediction
Hi all, I have a strange problem when applying RF in R. I have a set of variables with which I obtain an AUC of 0.67. I do have a second set of variables that have an AUC of 0.57. When I merge the first and second set of variables, the AUC becomes 0.64. I would expect the prediction to become better as I add variables that do have some predictive power? This is even more strange as the AUC
2007 Aug 24
2
Variable Importance - Random Forest
Hello, I am trying to explore the use of random forests for classification and am certain about the interpretation of the importance measurements. When having the option "importance = T" in the randomForest call, the resulting 'importance' element matrix has four columns with the following headings: 0 - mean raw importance score of variable x for class 0 (where
2007 Jan 29
3
comparing random forests and classification trees
Hi, I have done an analysis using 'rpart' to construct a Classification Tree. I am wanting to retain the output in tree form so that it is easily interpretable. However, I am wanting to compare the 'accuracy' of the tree to a Random Forest to estimate how much predictive ability is lost by using one simple tree. My understanding is that the error automatically displayed by the two
2013 Oct 18
3
fraud detection
hello everyone. i am concerned about security to the PBX and i would like to discuss different fraud detection methods. Apart from making everything to secure the PBX (latest patches, iptables, firewalls, no outside users, strongs passwds,...) i would like to find out if there are any fraud detection techniques. As for my setup i do have a PBX running asterisk 11.4 and it has 3 sip trunks (over
2004 Mar 02
1
some question regarding random forest
Hi, I had two questions regarding random forests for regression. 1) I have read the original paper by Breiman as well as a paper dicussing an application of random forests and it appears that the one of the nice features of this technique is good predictive ability. However I have some data with which I have generated a linear model using lm(). I can get an RMS error of 0.43 and an R^2 of