Displaying 20 results from an estimated 30000 matches similar to: "mboost partial contribution plots"
2010 Feb 02
0
Major update: mboost 2.0-0 released
Dear useRs,
we are happy to announce the release of mboost 2.0-0 on CRAN:
http://cran.r-project.org/package=mboost
This version contains major updates and changes to the implementation of
the main algorithm. Some slight changes to the user-interface where
necessary. Please consult the manual and the list of CHANGES below.
The package 'mboost' (Model-based Boosting) implements
2010 Feb 02
0
Major update: mboost 2.0-0 released
Dear useRs,
we are happy to announce the release of mboost 2.0-0 on CRAN:
http://cran.r-project.org/package=mboost
This version contains major updates and changes to the implementation of
the main algorithm. Some slight changes to the user-interface where
necessary. Please consult the manual and the list of CHANGES below.
The package 'mboost' (Model-based Boosting) implements
2010 Feb 03
0
mboost: how to implement cost-sensitive boosting family
mboost contains a blackboost method to build tree-based boosting models. I tried to write my own "cost-sensitive" ada family. But obviously my understanding to implement ngradient, loss, and offset functions is not right. I would greatly appreciate if anyone can help me out, or show me how to write a cost-sensitive family, thanks!
Follows are some families I wrote
ngradient <-
2012 Nov 04
1
blackboost (mboost package) function leads to non-reclaimable memory usage
Dear all,
I am puzzled by R's memory usage when calling the blackboost function from
package mboost to estimate a Gradient boosting model on a simulated dataset
with 20 correlated variables and 100,000 obs. The blackboost object created
by the function is only 15.3Mb, but R's memory usage increases by about
3.9Gb during the estimation of the model and the memory is not released even
after
2010 Mar 19
0
mboost: Interpreting coefficients from glmboost if center=TRUE
Sorry for the tardy reply but I just found your posting incidentally
today. To make long things short:
You are right about the centering. We forgot to correct the intercept if
center = TRUE. We lately found the problem ourself and fixed it in the
current version (mboost 2.0-3). However the problem only occurred if you
extracted the coefficients. As the intercept is rarely interpretable we
2010 Sep 21
1
package gbm, predict.gbm with offset
Dear all,
the help file for predict.gbm states that "The predictions from gbm do not
include the offset term. The user may add the value of the offset to the
predicted value if desired." I am just not sure how exactly, especially for
a Poisson model, where I believe the offset is multiplicative ?
For example:
library(MASS)
fit1 <- glm(Claims ~ District + Group + Age +
2008 Nov 12
2
apply() just loops ?
Dear R users,
I have been diligently using the apply() family in order to avoid explicit
for loops and speed up computation.
However, when I finally inspected the source code for apply, it appears that
the core computation is a simple loop as well.
What am I missing ? Why the often found advice to use apply() instead of
loops and the actually observed empirical speedups on many tasks ?
Thanks in
2012 Jul 23
1
mboost vs gbm
I'm attempting to fit boosted regression trees to a censored response using
IPCW weighting. I've implemented this through two libraries, mboost and
gbm, which I believe should yield models that would perform comparably.
This, however, is not the case - mboost performs much better. This seems
odd. This issue is meaningful since the output of this regression needs to
be implemented in a
2008 Apr 21
1
estimate of overdispersion with glm.nb
Dear R users,
I am trying to fully understand the difference between estimating
overdispersion with glm.nb() from MASS compared to glm(..., family =
quasipoisson).
It seems that (i) the coefficient estimates are different and also (ii) the
summary() method for glm.nb suggests that overdispersion is taken to be one:
"Dispersion parameter for Negative Binomial(0.9695) family taken to be
2010 Oct 20
1
problem with predict(mboost,...)
Hi,
I use a mboost model to predict my dependent variable on new data. I get the following warning message:
In bs(mf[[i]], knots = args$knots[[i]]$knots, degree = args$degree, :
some 'x' values beyond boundary knots may cause ill-conditioned bases
The new predicted values are partly negative although the variable in the training data ranges from 3 to 8 on a numeric scale. In order to
2008 Apr 17
2
pnbinom.c qnorm.c
Dear R users,
I was wondering from where I could get the C source code to compute
pnbinom() and qnorm() ?
(I would use R in batch mode but I find the startup time prohibitive, unless
there is a way to speed it up)
I searched the Web and it clearly is part of the R distribution, I just
don't know how to extract them.
Thanking you !
Markus Loecher
Princeton, NJ
[[alternative HTML version
2016 Nov 14
0
Major update of package actuar
Dear useRs,
I'm happy to announce a substantial update of package actuar that bumps the version number to 2.0-0. This release focuses on additional support for continuous and discrete distributions, new functions to simulate data from compound models and mixtures, and revised and improved documentation.
A slightly shortened version of the NEWS file follows:
NEW FEATURES
? Support for the
2016 Nov 14
0
Major update of package actuar
Dear useRs,
I'm happy to announce a substantial update of package actuar that bumps the version number to 2.0-0. This release focuses on additional support for continuous and discrete distributions, new functions to simulate data from compound models and mixtures, and revised and improved documentation.
A slightly shortened version of the NEWS file follows:
NEW FEATURES
? Support for the
2011 Feb 08
0
Update: googleVis 0.2.4 - Using the Google Visualisation API with R
Hi all,
Version 0.2.4 of the googVis package has been released on CRAN and
will be available from your local CRAN mirror soon.
googleVis provides an interface between R and the Google Visualisation API.
The functions of the package allow users to visualise data stored in R with the
Google Visualisation API without uploading their data to Google
Since the last version a lot of work has been
2011 Feb 08
0
Update: googleVis 0.2.4 - Using the Google Visualisation API with R
Hi all,
Version 0.2.4 of the googVis package has been released on CRAN and
will be available from your local CRAN mirror soon.
googleVis provides an interface between R and the Google Visualisation API.
The functions of the package allow users to visualise data stored in R with the
Google Visualisation API without uploading their data to Google
Since the last version a lot of work has been
2006 May 27
2
boosting - second posting
Hi
I am using boosting for a classification and prediction problem.
For some reason it is giving me an outcome that doesn't fall between 0
and 1 for the predictions. I have tried type="response" but it made no
difference.
Can anyone see what I am doing wrong?
Screen output shown below:
> boost.model <- gbm(as.factor(train$simNuance) ~ ., # formula
+
2020 Jan 19
0
rpois(9, 1e10)
Floor (maybe round) of non-negative numerics, though. Poisson should never
have anything after decimal.
Still think it?s worth allowing long long for R64 bit, just for purity
sake.
Avi
On Sun, Jan 19, 2020 at 4:38 PM Spencer Graves <spencer.graves at prodsyse.com>
wrote:
>
>
> On 2020-01-19 13:01, Avraham Adler wrote:
>
> Crazy thought, but being that a sum of Poissons is
2010 Feb 07
1
mboost: Interpreting coefficients from glmboost if center=TRUE
I'm running R 2.10.1 with mboost 2.0 in order to build predictive
models . I am performing prediction on a binomial outcome, using a
linear function (glmboost). However, I am running into some confusion
regarding centering. (I am not aware of an mboost-specific mailing
list, so if the main R list is not the right place for this topic,
please let me know.)
The boost_control() function allows
2010 Jul 28
2
Out-of-sample predictions with boosting model
Hi UseRs -
I am new to R, and could use some help making out-of-sample predictions
using a boosting model (the mboost command). The issue is complicated by the
fact that I have panel data (time by country), and am estimating the model
separately for each country. FYI, this is monthly data and I have 1986m1 -
2009m12 for 9 countries.
To give you a flavor of what I am doing, here is a simple
2006 May 25
0
boosting
Hi
I am using boosting for a classification and prediction problem.
For some reason it is giving me an outcome that doesn't fall between 0
and 1 for the predictions. I have tried type="response" but it made no
difference.
Can anyone see what I am doing wrong?
Screen output shown below:
> boost.model <- gbm(as.factor(train$simNuance) ~ ., # formula
+