Displaying 20 results from an estimated 20000 matches similar to: "Packages for multi-class classification with boosting"
2006 Mar 12
1
boosting for multi-class classification
Hi List,
I can't seem to find a package that implements boosting for
multi-class classification.
Does such a package exist?
2006 May 27
2
boosting - second posting
Hi
I am using boosting for a classification and prediction problem.
For some reason it is giving me an outcome that doesn't fall between 0
and 1 for the predictions. I have tried type="response" but it made no
difference.
Can anyone see what I am doing wrong?
Screen output shown below:
> boost.model <- gbm(as.factor(train$simNuance) ~ ., # formula
+
2006 May 25
0
boosting
Hi
I am using boosting for a classification and prediction problem.
For some reason it is giving me an outcome that doesn't fall between 0
and 1 for the predictions. I have tried type="response" but it made no
difference.
Can anyone see what I am doing wrong?
Screen output shown below:
> boost.model <- gbm(as.factor(train$simNuance) ~ ., # formula
+
2009 Apr 07
0
gbm for multi-class problems
Dear List,
I´m working on a classification problem. My response has 60 levels.
I`m very interested in boosted trees like AdaBoost or gradient boosting machine as implemented in the package "gbm". Unfortunately gbm is only applicable for 2-class problems.
Is anybody out there who can help me? Is there a way to use gbm() for multi-class problems? Maybe there is a way to transform my
2005 Jul 12
1
SOS Boosting
Hi,
I am trying to implement the Adaboost.M1. algorithm as described in
"The Elements of Statistical Learning" p.301
I don't use Dtettling 's library "boost" because :
- I don't understande the difference beetween Logitboost and L2boost
- I 'd like to use larger trees than stumps.
By using option weights set to (1/n, 1/n, ..., 1/n) in rpart or tree
2005 Mar 21
1
Convert numeric to class
Dear all,
I have a script about iteration classification, like this below
data(iris)
N <- 5
ir.tr.iter <- vector('list',N)
ir.tr <- vector('list',N)
for (j in 1:N) {
ir.tr[[j]] <- rpart(Species ~., data=iris)
ir.tr.iter[j] <- ir.tr[[j]]$frame
result <- list(ir.tr=ir.tr, ir.tr.iter=ir.tr.iter)
}
as.data.frame(as.matrix(ir.tr.iter))
2005 Dec 30
4
Count or summary data
Dear all,
I want to summary and count my data something like
> te.Ce
[,1] [,2]
[1,] -1 0.05
[2,] 1 0.05
[3,] 1 0.00
[4,] 0 0.05
[5,] -1 0.00
[6,] 0 0.10
[7,] 1 0.10
[8,] -1 0.00
[9,] -1 0.10
[10,] 0 0.05
[11,] 0 0.10
[12,] -1 0.10
[13,] 1 0.00
[14,] -1 0.05
[15,] 1 0.00
How could I count (summary) all my data which I need the result like
2009 Aug 26
0
Doubt about adaboost
Hello,
I performed a boosting analisis with adabag package to obtain a classification tree with the following set of commands:
Tesis.boost <- adaboost.M1(Captura~., data=Tesis2, mfinal=2)
> arb<-Tesis.boost$tree[[1]]
> post(arb, file ="")
> post(arb, file ="",title= "Arbol 1")
I would like to know the meanning of the numbers that appeared in the
2005 Mar 21
5
Read a dataset with different lengths
Dear useR again,
How can I read a dataset if lines in dataset did not have same
elements (have different lengths), For example:
1 2, 4, 16, 1, 1, 3, 1, 1, 15, 5, 1, 1, 14, 1, 1
2 2, 13, 5, 1, 1, 3, 1, 1, 15, 5, 1, 1, 14, 1, 1
3 4, 5, 11, 1, 1, 6, 1, 1, 5, 14, 1, 1, 15, 1, 1
4 2, 5, 9, 1, 1, 14, 1, 1, 8, 16, 1, 1, 13, 1, 1
5 3, 7, 14, 1,
2003 Jul 14
0
package announcement: Generalized Boosted Models (gbm)
Generalized Boosted Models (gbm)
This package implements extensions to Y. Freund and R. Schapire's AdaBoost
algorithm and J. Friedman's gradient boosting machine (aka multivariate
adaptive regression trees, MART). It includes regression methods for least
squares, absolute loss, logistic, Poisson, Cox proportional hazards/partial
likelihood, and the AdaBoost exponential loss. It handles
2003 Jul 14
0
package announcement: Generalized Boosted Models (gbm)
Generalized Boosted Models (gbm)
This package implements extensions to Y. Freund and R. Schapire's AdaBoost
algorithm and J. Friedman's gradient boosting machine (aka multivariate
adaptive regression trees, MART). It includes regression methods for least
squares, absolute loss, logistic, Poisson, Cox proportional hazards/partial
likelihood, and the AdaBoost exponential loss. It handles
2017 Dec 14
0
Distributions for gbm models
On page 409 of "Applied Predictive Modeling" by Max Kuhn, it states
that the gbm function can accomodate only two class problems when
referring to the distribution parameter.
>From gbm help re: the distribution parameter:
Currently available options are "gaussian" (squared error),
"laplace" (absolute loss), "tdist" (t-distribution
2009 Jun 19
2
good boosting tutorial and package in R?
Hi all,
Could you please give me some pointers about what's the best boosting
package in R currently?
in terms of classification accuracy?
And any pointers about tutorials and study-materials to curb the
learning curve will be greatly appreciated!
Thank you!
p.s. Does anybody happen to know Boosting implemented in other
language such as Matlab?
Are they good in terms of accuracy?
What
2005 Aug 15
2
How to get a list work in RData file
Dear R-Helper,
I want to know how I get a list work which I saved in RData file. For
example,
> test.xy <- function(x,y) {
+ xy <- x+y
+ xy
+ }
>
> xyadd <- test.xy(x=2, y=3)
> xyadd
[1] 5
> x1 <- c(2,43,60,8)
> y1 <- c(91,7,5,30)
>
> xyadd1 <- test.xy(x=x1, y=y1)
> xyadd1
[1] 93 50 65 38
> save(list = ls(all=TRUE), file =
2017 Jul 18
1
Help-Multi class classification for large datasets
Hai all,
We are working on Multi-class Classification. Currently up to 1.1 million
records Ranger package in R is able to handle. Training time on 128 GB RAM
is 12 days, which is not a practically feasible method to proceed further.
In future we will have dataset of dimension 10 million records, we are in
search for a package or framework which can handle 10 million records with
at least 12000
2002 Jun 23
2
AdaBoost for R
I'm going to implement AdaBoost algorithm in R. Just wanted to ensure
that there is no implementation of any boosting algorithm in R... don't
want
to reinvent the wheel...
-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-
r-devel mailing list -- Read http://www.ci.tuwien.ac.at/~hornik/R/R-FAQ.html
Send "info", "help", or
2005 Jan 25
0
Collapsing solution to the question discussed above: Re: multi-class classification using rpart
You could break your 3 class problem into several (2 or 3) 2 class problems,
and then use Andy's suggestion (see the CART book). There are several ways
to break the problem into 2 class problems, and several ways to combine the
resulting classifiers. Tom Dietterich, Jerry Friedman, Trevor Hastie and Rob
Tibshirani, among others, have articles on the question, in places like
Annals of
2015 Dec 01
0
[RFC PATCH 5/5] clk: allow boosting only when NvBoost is set
0: disable boosting (cap to base clock from the vbios)
1: boost only to boost clock from the vbios
2: boost to max clock available
---
drm/nouveau/include/nvkm/subdev/clk.h | 10 +++++++++-
drm/nouveau/nvkm/subdev/clk/base.c | 17 ++++++++++++++++-
drm/nouveau/nvkm/subdev/clk/gf100.c | 2 +-
drm/nouveau/nvkm/subdev/clk/gk104.c | 2 +-
4 files changed, 27 insertions(+), 4 deletions(-)
2004 Sep 27
1
random discrete from the first tuple
Dear UseR
I have a dataset, for instance x1(A, B, C), x2(M,F), x3(X1,X2,X3,X4)
and x4(W,F,P). I want to make a small dataset with the random tuple. I
know package e1071 can handle a random discrete,
> library(e1071)
> x1 <- rdiscrete(6, c(2,2,2), c("A","B","C"))
> x2 <- rdiscrete(6, c(3,3), c("M","F"))
> x3 <-
2015 Dec 02
0
[RFC PATCH 5/5] clk: allow boosting only when NvBoost is set
> Pierre Moreau <pierre.morrow at free.fr> hat am 2. Dezember 2015 um 02:26
> geschrieben:
>
> Hi Karol,
>
> I have some comments below.
>
> On 05:42 PM - Dec 01 2015, Karol Herbst wrote:
> > 0: disable boosting (cap to base clock from the vbios)
> > 1: boost only to boost clock from the vbios
> > 2: boost to max clock available
> > ---
>