Displaying 20 results from an estimated 1000 matches similar to: "package announcement: Generalized Boosted Models (gbm)"
2010 Apr 26
3
R.GBM package
HI, Dear Greg,
I AM A NEW to GBM package. Can boosting decision tree be implemented in
'gbm' package? Or 'gbm' can only be used for regression?
IF can, DO I need to combine the rpart and gbm command?
Thanks so much!
--
Sincerely,
Changbin
--
[[alternative HTML version deleted]]
2009 Apr 07
0
gbm for multi-class problems
Dear List,
I´m working on a classification problem. My response has 60 levels.
I`m very interested in boosted trees like AdaBoost or gradient boosting machine as implemented in the package "gbm". Unfortunately gbm is only applicable for 2-class problems.
Is anybody out there who can help me? Is there a way to use gbm() for multi-class problems? Maybe there is a way to transform my
2006 Sep 18
0
Propensity score modeling using machine learning methods. WAS: RE: LARS for generalized linear models
There may be benefits to having a machine learning method that
explicitly targets covariate balance. We have experimented with
optimizing the weights directly to obtain the best covariate balance,
but got some strange solutions for simple cases that made us wary of
such methods.
Machine learning methods that yield calibrated probability estimates
should do well (e.g. those that optimize the
2010 May 01
1
bag.fraction in gbm package
Hi, Dear Greg,
Sorry to bother you again.
I have several questions about the 'gbm' package.
if the train.fraction is less than 1 (ie. 0.5) , then the* first* 50% will
be used to fit the model, the other 50% can be used to estimate the
performance.
if bag.fraction is 0.5, then gbm use the* random* 50% of the data to fit the
model, and the other 50% data is used to estimate the
2013 Jun 23
1
Which is the final model for a Boosted Regression Trees (GBM)?
Hi R User,
I was trying to find a final model in the following example by using the Boosted regression trees (GBM). The program gives the fitted values but I wanted to calculate the fitted value by hand to understand in depth. Would you give moe some hints on what is the final model for this example?
Thanks
KG
-------
The following script I used
#-----------------------
library(dismo)
2017 Dec 14
0
Distributions for gbm models
On page 409 of "Applied Predictive Modeling" by Max Kuhn, it states
that the gbm function can accomodate only two class problems when
referring to the distribution parameter.
>From gbm help re: the distribution parameter:
Currently available options are "gaussian" (squared error),
"laplace" (absolute loss), "tdist" (t-distribution
2005 Feb 18
2
gbm
Hi, there:
I am always experiencing the scalability of some R packages. This
time, I am trying gbm to do adaboosting on my project. Initially I
tried to grow trees by using rpart on a dataset with 200 variables and
30,000 observations. Now, I am thinking if I can apply adaboosting on
it.
I am wondering if here is anyone who did a similar thing before and
can provide some sample codes. Also any
2006 May 27
2
boosting - second posting
Hi
I am using boosting for a classification and prediction problem.
For some reason it is giving me an outcome that doesn't fall between 0
and 1 for the predictions. I have tried type="response" but it made no
difference.
Can anyone see what I am doing wrong?
Screen output shown below:
> boost.model <- gbm(as.factor(train$simNuance) ~ ., # formula
+
2010 Feb 28
1
Gradient Boosting Trees with correlated predictors in gbm
Dear R users,
I’m trying to understand how correlated predictors impact the Relative
Importance measure in Stochastic Boosting Trees (J. Friedman). As Friedman
described “ …with single decision trees (referring to Brieman’s CART
algorithm), the relative importance measure is augmented by a strategy
involving surrogate splits intended to uncover the masking of influential
variables by others
2010 Sep 21
1
package gbm, predict.gbm with offset
Dear all,
the help file for predict.gbm states that "The predictions from gbm do not
include the offset term. The user may add the value of the offset to the
predicted value if desired." I am just not sure how exactly, especially for
a Poisson model, where I believe the offset is multiplicative ?
For example:
library(MASS)
fit1 <- glm(Claims ~ District + Group + Age +
2009 Oct 30
1
possible memory leak in predict.gbm(), package gbm ?
Dear gbm users,
When running predict.gbm() on a "large" dataset (150,000 rows, 300 columns,
500 trees), I notice that the memory used by R grows beyond reasonable
limits. My 14GB of RAM are often not sufficient. I am interpreting this as a
memory leak since there should be no reason to expand memory needs once the
data are loaded and passed to predict.gbm() ?
Running R version 2.9.2 on
2012 Sep 16
2
Where is the R configuration file or how to override R compilers
I have a question about how one can modify or override the compilers
that R uses for package installations? Or if perhaps this configuration
is in some editable file somewhere.
Initially I built the version of R 2.15.1 on Solaris SPARC (virtual T4),
but found out the build was done as 32 bit. After some research, I
found that the pre-compiled GCC version I had only allowed for 32 bit.
I wanted
2011 May 24
1
gbm package: plotting a single tree
Hello,
I'm not sure if Im posting this on the right place, my apologies if not.
I'm using the package gbm to generate boosted trees models, and was
wondering if there is a simple way of getting a graphical output for a
single tree of the sequence. I know the function "pretty.gbm.tree" can be
used to print information for a single tree, but I've been unable to find a
way to
2012 Jul 23
1
mboost vs gbm
I'm attempting to fit boosted regression trees to a censored response using
IPCW weighting. I've implemented this through two libraries, mboost and
gbm, which I believe should yield models that would perform comparably.
This, however, is not the case - mboost performs much better. This seems
odd. This issue is meaningful since the output of this regression needs to
be implemented in a
2012 Dec 12
1
extracting splitting rules from GBM
I extracting splitting rules from Greg Ridgeway's GBM 1.6-3.2 in R 2.15.2, so I can run classification in a production system outside of R. ?I have it working and verified for a dummy data set with all variable types (numeric, factor, ordered) and missing values, but in the titanic survivors data set the splitting rule for factors does not make sense. ?The attached code and log below explains
2006 May 25
0
boosting
Hi
I am using boosting for a classification and prediction problem.
For some reason it is giving me an outcome that doesn't fall between 0
and 1 for the predictions. I have tried type="response" but it made no
difference.
Can anyone see what I am doing wrong?
Screen output shown below:
> boost.model <- gbm(as.factor(train$simNuance) ~ ., # formula
+
2009 Apr 14
3
Problem cross-compiling on Ubuntu
I'm using Ubuntu 8.10 (Intrepid Ibex) and R 2.7.1.
I've built a package from source (a modified version of gbm) and it
contains some C++ code. I now want to cross-compile it to get a
Windows version.
I installed R using
sudo apt-get update
sudo apt-get install r-base
sudo apt-get install r-base-dev
So far as I can tell, I've also followed all the instructions in the
guide
2013 Mar 24
3
Parallelizing GBM
Dear All,
I am far from being a guru about parallel programming.
Most of the time, I rely or randomForest for data mining large datasets.
I would like to give a try also to the gradient boosted methods in GBM,
but I have a need for parallelization.
I normally rely on gbm.fit for speed reasons, and I usually call it this
way
gbm_model <- gbm.fit(trainRF,prices_train,
offset = NULL,
misc =
2009 Dec 14
0
GBM package: Extract coefficients
I am using the gbm package for generalized boosted regression models,
and would like to be able to extract the coefficients produced for
storage in a database.
I am already using R to automatically generate formulas that I can
export to a database and store. For example, I have been using Dr.
Harrell's lrm package to perform logistic regression, e.g.:
output <-
2011 Feb 26
2
Reproducibility issue in gbm (32 vs 64 bit)
Dear List,
The gbm package on Win 7 produces different results for the
relative importance of input variables in R 32-bit relative to R 64-bit. Any
idea why? Any idea which one is correct?
Based on this example, it looks like the relative importance of 2 perfectly
correlated predictors is "diluted" by half in 32-bit, whereas in 64-bit, one
of these predictors gets all the importance