similar to: some question regarding random forest

Displaying 20 results from an estimated 6000 matches similar to: "some question regarding random forest"

2003 Sep 16
1
simplifying randomForest(s)
Dear All, I have been using the randomForest package for a couple of difficult prediction problems (which also share p >> n). The performance is good, but since all the variables in the data set are used, interpretation of what is going on is not easy, even after looking at variable importance as produced by the randomForest run. I have tried a simple "variable selection"
2011 Feb 15
1
[slightly OT] predict.randomForest and type=”prob”
Dear all , I would like to use the function randomForest to predict the probability of relocation failure of a GPS collar as a function of several environmental variables x (both factor and numeric: slope, vegetation, etc.) on a given area. The response variable y is thus success (0)/failure(1) of the relocation, and the sampling unit is the pixel of a raster map. My aim is to build a map
2003 Aug 05
1
na.action in randomForest --- Summary
A few days ago I asked whether there were options other than na.action=na.fail for the R port of Breiman?s randomForest; the function?s help page did not say anything about other options. I have since discovered that a pdf document called ?The randomForest Package? and made available by Andy Liaw (who made the tool available in R---thank you) does discuss an option. It is an implementation of
2007 Aug 24
2
Variable Importance - Random Forest
Hello, I am trying to explore the use of random forests for classification and am certain about the interpretation of the importance measurements. When having the option "importance = T" in the randomForest call, the resulting 'importance' element matrix has four columns with the following headings: 0 - mean raw importance score of variable x for class 0 (where
2011 Oct 10
1
pmml for random forest & rules
Hi, I am having some trouble using R 2.13.1 for generating a pmml object of of class "c('randomForest.formula', 'randomForest')" I see that these methods are available: > methods(pmml) [1] pmml.coxph* pmml.hclust* pmml.itemsets* pmml.kmeans* pmml.ksvm* pmml.lm* pmml.multinom* pmml.nnet* pmml.rpart* [10] pmml.rsf* pmml.rules* pmml.survreg*
2006 Jul 26
3
memory problems when combining randomForests
Dear all, I am trying to train a randomForest using all my control data (12,000 cases, ~ 20 explanatory variables, 2 classes). Because of memory constraints, I have split my data into 7 subsets and trained a randomForest for each, hoping that using combine() afterwards would solve the memory issue. Unfortunately, combine() still runs out of memory. Is there anything else I can do? (I am not using
2002 Apr 02
2
random forests for R
Hi all, There is now a package available on CRAN that provides an R interface to Leo Breiman's random forest classifier. Basically, random forest does the following: 1. Select ntree, the number of trees to grow, and mtry, a number no larger than number of variables. 2. For i = 1 to ntree: 3. Draw a bootstrap sample from the data. Call those not in the bootstrap sample the
2002 Apr 02
2
random forests for R
Hi all, There is now a package available on CRAN that provides an R interface to Leo Breiman's random forest classifier. Basically, random forest does the following: 1. Select ntree, the number of trees to grow, and mtry, a number no larger than number of variables. 2. For i = 1 to ntree: 3. Draw a bootstrap sample from the data. Call those not in the bootstrap sample the
2008 Jul 22
2
randomForest Tutorial
I am new to R and I'd like to use the randomForest package for my thesis (identifying important variables for more detailed analysis with other software). I have found extremely well written and helpful information on the usage of R. Unfortunately it seems to be very difficult to find similarly detailed tutorials for randomForest, and I just can't get it work with the information on
2006 Mar 29
2
missing value replacement for test data in random forest
Hi, In R, how to do missing value replacement for test data in randome forest in the way Breiman decribed. thanks in advance iris
2011 Sep 13
1
class weights with Random Forest
Hi All, I am looking for a reference that explains how the randomForest function in the randomForest package uses the classwt parameter. Here: http://tolstoy.newcastle.edu.au/R/e4/help/08/05/12088.html Andy Liaw suggests not using classwt. And according to: http://r.789695.n4.nabble.com/R-help-with-RandomForest-classwt-option-td817149.html it has "not been implemented" as of 2007.
2002 Sep 25
5
CART vs. Random Forest
According to Dr. Breiman, the RF should be more accurate method than a single tree. However, the performance of each method seems to depend on the proprotion of outcome variable in my case. My data set is a typical classification problem (predict bad guys). When I ran both of them with different proportion of outcome variables(there's a criterion to measure the degree of bad behavior), I
2009 Apr 13
2
Random Forests Variable Importance Question
I am trying to use the random forests package for classification in R. The Variable Importance Measures listed are: -mean raw importance score of variable x for class 0 -mean raw importance score of variable x for class 1 -MeanDecreaseAccuracy -MeanDecreaseGini Now I know what these "mean" as in I know their definitions. What I want to know is how to use them. What I am trying to
2004 Apr 18
2
outliers using Random Forest
Hello, Does anybody know if the outscale option of randomForest yields the standarized version of the outlier measure for each case? or the results are only the raw values. Also I have notice that this measure presents very high variability. I mean if I repeat the experiment I am getting very different values for this measure and it is hard to flag the outliers. This does not happen with two other
2012 Dec 03
2
Different results from random.Forest with test option and using predict function
Hello R Gurus, I am perplexed by the different results I obtained when I ran code like this: set.seed(100) test1<-randomForest(BinaryY~., data=Xvars, trees=51, mtry=5, seed=200) predict(test1, newdata=cbind(NewBinaryY, NewXs), type="response") and this code: set.seed(100) test2<-randomForest(BinaryY~., data=Xvars, trees=51, mtry=5, seed=200, xtest=NewXs, ytest=NewBinarY) The
2009 Jun 24
1
Random Forest Variable Importance Interpretation
Hi I am trying to explore the use of random forests for regression to identify the important environmental/microclimate variables involved in predicting the abundance of a species in different habitats, there are approx 40 variable and between 200 and 500 data points depending on the dataset. I have successfully used the randomForest package to conduct the analysis and looked at the %IncMSE
2017 Jun 03
2
CV en R
?Hola, Puedes ver aquí un ejemplo de cómo comparar varios modelos usando "caret". https://stackoverflow.com/questions/14800021/statistics-of-prediction-for-multiple-models-with-caret O mejor en el propio manual de "caret", en esta sección: https://topepo.github.io/caret/model-training-and-tuning.html#exploring-and-comparing-resampling-distributions Y como recomendación te
2017 Jun 02
2
CV en R
No, llega un momento en el que más árboles no te supone mejoría, e incluso funciona peor. Que funcione peor lo atribuyo al ruido, porque en teoría no tiene mucho sentido, la verdad... Pero no he probado a coger más árboles de los "necesarios". Lo probaré… Un saludo De: Jesús Para Fernández [mailto:j.para.fernandez en hotmail.com] Enviado el: viernes, 02 de junio de 2017 14:54
2003 Nov 24
2
Questions on Random Forest
Hi, everyone, I am a newbie on R. Now I want to do image pixel classification by random forest. But I has not a clear understanding on random forest. Here is some question: As for an image, for example its size is 512x512 and has only one variable -- gray level. The histogram of the image looks like mixture Gaussian Model, say Gauss distribution (u1,sigma1), (u2,sigma2),(u3,sigma3). And a
2017 Jun 04
2
CV en R
Si nos dices el tipo de problema que estás intentando solucionar y el tamaño del dataset podemos recomendarte algo más. En tu pseudo-código mezclas algoritmos supervisados y no-supervisados. Además de ranger, daría alguna oportunidad a "gbm" o como no a "xgboost". Y éstos los probaría dentro de H2O. Saludos, Carlos Ortega www.qualityexcellence.es El 4 de junio de 2017,