Displaying 20 results from an estimated 6000 matches similar to: "some question regarding random forest"
2003 Sep 16
1
simplifying randomForest(s)
Dear All,
I have been using the randomForest package for a couple of difficult
prediction problems (which also share p >> n). The performance is good, but
since all the variables in the data set are used, interpretation of what is
going on is not easy, even after looking at variable importance as produced
by the randomForest run.
I have tried a simple "variable selection"
2011 Feb 15
1
[slightly OT] predict.randomForest and type=”prob”
Dear all ,
I would like to use the function randomForest to predict the probability
of relocation failure of a GPS collar as a function of several
environmental variables x (both factor and numeric: slope, vegetation,
etc.) on a given area. The response variable y is thus success
(0)/failure(1) of the relocation, and the sampling unit is the pixel of
a raster map. My aim is to build a map
2003 Aug 05
1
na.action in randomForest --- Summary
A few days ago I asked whether there were options other than
na.action=na.fail for the R port of Breiman?s randomForest; the function?s
help page did not say anything about other options.
I have since discovered that a pdf document called ?The randomForest
Package? and made available by Andy Liaw (who made the tool available in
R---thank you) does discuss an option. It is an implementation of
2007 Aug 24
2
Variable Importance - Random Forest
Hello,
I am trying to explore the use of random forests for classification and
am certain about the interpretation of the importance measurements.
When having the option "importance = T" in the randomForest call, the
resulting 'importance' element matrix has four columns with the
following headings:
0 - mean raw importance score of variable x for class 0 (where
2011 Oct 10
1
pmml for random forest & rules
Hi,
I am having some trouble using R 2.13.1 for generating a pmml object
of of class "c('randomForest.formula', 'randomForest')"
I see that these methods are available:
> methods(pmml)
[1] pmml.coxph* pmml.hclust* pmml.itemsets* pmml.kmeans*
pmml.ksvm* pmml.lm* pmml.multinom* pmml.nnet*
pmml.rpart*
[10] pmml.rsf* pmml.rules* pmml.survreg*
2006 Jul 26
3
memory problems when combining randomForests
Dear all,
I am trying to train a randomForest using all my control data (12,000 cases, ~
20 explanatory variables, 2 classes). Because of memory constraints, I have
split my data into 7 subsets and trained a randomForest for each, hoping that
using combine() afterwards would solve the memory issue. Unfortunately,
combine() still runs out of memory. Is there anything else I can do? (I am not
using
2002 Apr 02
2
random forests for R
Hi all,
There is now a package available on CRAN that provides an R interface to Leo
Breiman's random forest classifier.
Basically, random forest does the following:
1. Select ntree, the number of trees to grow, and mtry, a number no larger
than number of variables.
2. For i = 1 to ntree:
3. Draw a bootstrap sample from the data. Call those not in the bootstrap
sample the
2002 Apr 02
2
random forests for R
Hi all,
There is now a package available on CRAN that provides an R interface to Leo
Breiman's random forest classifier.
Basically, random forest does the following:
1. Select ntree, the number of trees to grow, and mtry, a number no larger
than number of variables.
2. For i = 1 to ntree:
3. Draw a bootstrap sample from the data. Call those not in the bootstrap
sample the
2008 Jul 22
2
randomForest Tutorial
I am new to R and I'd like to use the randomForest package for my thesis
(identifying important variables for more detailed analysis with other
software). I have found extremely well written and helpful information on
the usage of R.
Unfortunately it seems to be very difficult to find similarly detailed
tutorials for randomForest, and I just can't get it work with the
information on
2006 Mar 29
2
missing value replacement for test data in random forest
Hi,
In R, how to do missing value replacement for test data in randome forest in the way Breiman decribed.
thanks in advance
iris
2011 Sep 13
1
class weights with Random Forest
Hi All,
I am looking for a reference that explains how the randomForest function in
the randomForest package uses the classwt parameter. Here:
http://tolstoy.newcastle.edu.au/R/e4/help/08/05/12088.html
Andy Liaw suggests not using classwt. And according to:
http://r.789695.n4.nabble.com/R-help-with-RandomForest-classwt-option-td817149.html
it has "not been implemented" as of 2007.
2002 Sep 25
5
CART vs. Random Forest
According to Dr. Breiman, the RF should be more accurate
method than a single tree. However, the performance of each
method seems to depend on the proprotion of outcome variable
in my case. My data set is a typical classification problem
(predict bad guys). When I ran both of them with different
proportion of outcome variables(there's a criterion to measure
the degree of bad behavior), I
2009 Apr 13
2
Random Forests Variable Importance Question
I am trying to use the random forests package for classification in R.
The Variable Importance Measures listed are:
-mean raw importance score of variable x for class 0
-mean raw importance score of variable x for class 1
-MeanDecreaseAccuracy
-MeanDecreaseGini
Now I know what these "mean" as in I know their definitions. What I
want to know is how to use them.
What I am trying to
2004 Apr 18
2
outliers using Random Forest
Hello,
Does anybody know if the outscale option of randomForest yields the
standarized version of the outlier measure for each case? or the results
are only the raw values. Also I have notice that this measure presents
very high variability. I mean if I repeat the experiment I am getting very
different values for this measure and it is hard to flag the outliers.
This does not happen with two other
2012 Dec 03
2
Different results from random.Forest with test option and using predict function
Hello R Gurus,
I am perplexed by the different results I obtained when I ran code like
this:
set.seed(100)
test1<-randomForest(BinaryY~., data=Xvars, trees=51, mtry=5, seed=200)
predict(test1, newdata=cbind(NewBinaryY, NewXs), type="response")
and this code:
set.seed(100)
test2<-randomForest(BinaryY~., data=Xvars, trees=51, mtry=5, seed=200,
xtest=NewXs, ytest=NewBinarY)
The
2009 Jun 24
1
Random Forest Variable Importance Interpretation
Hi
I am trying to explore the use of random forests for regression to
identify the important environmental/microclimate variables involved in
predicting the abundance of a species in different habitats, there are
approx 40 variable and between 200 and 500 data points depending on the
dataset. I have successfully used the randomForest package to conduct
the analysis and looked at the %IncMSE
2017 Jun 03
2
CV en R
?Hola,
Puedes ver aquí un ejemplo de cómo comparar varios modelos usando "caret".
https://stackoverflow.com/questions/14800021/statistics-of-prediction-for-multiple-models-with-caret
O mejor en el propio manual de "caret", en esta sección:
https://topepo.github.io/caret/model-training-and-tuning.html#exploring-and-comparing-resampling-distributions
Y como recomendación te
2017 Jun 02
2
CV en R
No, llega un momento en el que más árboles no te supone mejoría, e incluso
funciona peor. Que funcione peor lo atribuyo al ruido, porque en teoría no
tiene mucho sentido, la verdad... Pero no he probado a coger más árboles de
los "necesarios". Lo probaré…
Un saludo
De: Jesús Para Fernández [mailto:j.para.fernandez en hotmail.com]
Enviado el: viernes, 02 de junio de 2017 14:54
2003 Nov 24
2
Questions on Random Forest
Hi, everyone,
I am a newbie on R. Now I want to do image pixel classification by random
forest. But I has not a clear understanding on random forest. Here is some
question:
As for an image, for example its size is 512x512 and has only one variable
-- gray level. The histogram of the image looks like mixture Gaussian Model,
say Gauss distribution (u1,sigma1), (u2,sigma2),(u3,sigma3). And a
2017 Jun 04
2
CV en R
Si nos dices el tipo de problema que estás intentando solucionar y el
tamaño del dataset podemos recomendarte algo más.
En tu pseudo-código mezclas algoritmos supervisados y no-supervisados.
Además de ranger, daría alguna oportunidad a "gbm" o como no a "xgboost". Y
éstos los probaría dentro de H2O.
Saludos,
Carlos Ortega
www.qualityexcellence.es
El 4 de junio de 2017,