similar to: Packages for multi-class classification with boosting

Displaying 20 results from an estimated 20000 matches similar to: "Packages for multi-class classification with boosting"

2006 Mar 12
1
boosting for multi-class classification
Hi List, I can't seem to find a package that implements boosting for multi-class classification. Does such a package exist?
2006 May 27
2
boosting - second posting
Hi I am using boosting for a classification and prediction problem. For some reason it is giving me an outcome that doesn't fall between 0 and 1 for the predictions. I have tried type="response" but it made no difference. Can anyone see what I am doing wrong? Screen output shown below: > boost.model <- gbm(as.factor(train$simNuance) ~ ., # formula +
2006 May 25
0
boosting
Hi I am using boosting for a classification and prediction problem. For some reason it is giving me an outcome that doesn't fall between 0 and 1 for the predictions. I have tried type="response" but it made no difference. Can anyone see what I am doing wrong? Screen output shown below: > boost.model <- gbm(as.factor(train$simNuance) ~ ., # formula +
2009 Apr 07
0
gbm for multi-class problems
Dear List, I´m working on a classification problem. My response has 60 levels. I`m very interested in boosted trees like AdaBoost or gradient boosting machine as implemented in the package "gbm". Unfortunately gbm is only applicable for 2-class problems. Is anybody out there who can help me? Is there a way to use gbm() for multi-class problems? Maybe there is a way to transform my
2005 Jul 12
1
SOS Boosting
Hi, I am trying to implement the Adaboost.M1. algorithm as described in "The Elements of Statistical Learning" p.301 I don't use Dtettling 's library "boost" because : - I don't understande the difference beetween Logitboost and L2boost - I 'd like to use larger trees than stumps. By using option weights set to (1/n, 1/n, ..., 1/n) in rpart or tree
2005 Mar 21
1
Convert numeric to class
Dear all, I have a script about iteration classification, like this below data(iris) N <- 5 ir.tr.iter <- vector('list',N) ir.tr <- vector('list',N) for (j in 1:N) { ir.tr[[j]] <- rpart(Species ~., data=iris) ir.tr.iter[j] <- ir.tr[[j]]$frame result <- list(ir.tr=ir.tr, ir.tr.iter=ir.tr.iter) } as.data.frame(as.matrix(ir.tr.iter))
2005 Dec 30
4
Count or summary data
Dear all, I want to summary and count my data something like > te.Ce [,1] [,2] [1,] -1 0.05 [2,] 1 0.05 [3,] 1 0.00 [4,] 0 0.05 [5,] -1 0.00 [6,] 0 0.10 [7,] 1 0.10 [8,] -1 0.00 [9,] -1 0.10 [10,] 0 0.05 [11,] 0 0.10 [12,] -1 0.10 [13,] 1 0.00 [14,] -1 0.05 [15,] 1 0.00 How could I count (summary) all my data which I need the result like
2009 Aug 26
0
Doubt about adaboost
Hello, I performed a boosting analisis with adabag package to obtain a classification tree with the following set of commands: Tesis.boost <- adaboost.M1(Captura~., data=Tesis2, mfinal=2) > arb<-Tesis.boost$tree[[1]] > post(arb, file ="") > post(arb, file ="",title= "Arbol 1") I would like to know the meanning of the numbers that appeared in the
2005 Mar 21
5
Read a dataset with different lengths
Dear useR again, How can I read a dataset if lines in dataset did not have same elements (have different lengths), For example: 1 2, 4, 16, 1, 1, 3, 1, 1, 15, 5, 1, 1, 14, 1, 1 2 2, 13, 5, 1, 1, 3, 1, 1, 15, 5, 1, 1, 14, 1, 1 3 4, 5, 11, 1, 1, 6, 1, 1, 5, 14, 1, 1, 15, 1, 1 4 2, 5, 9, 1, 1, 14, 1, 1, 8, 16, 1, 1, 13, 1, 1 5 3, 7, 14, 1,
2003 Jul 14
0
package announcement: Generalized Boosted Models (gbm)
Generalized Boosted Models (gbm) This package implements extensions to Y. Freund and R. Schapire's AdaBoost algorithm and J. Friedman's gradient boosting machine (aka multivariate adaptive regression trees, MART). It includes regression methods for least squares, absolute loss, logistic, Poisson, Cox proportional hazards/partial likelihood, and the AdaBoost exponential loss. It handles
2003 Jul 14
0
package announcement: Generalized Boosted Models (gbm)
Generalized Boosted Models (gbm) This package implements extensions to Y. Freund and R. Schapire's AdaBoost algorithm and J. Friedman's gradient boosting machine (aka multivariate adaptive regression trees, MART). It includes regression methods for least squares, absolute loss, logistic, Poisson, Cox proportional hazards/partial likelihood, and the AdaBoost exponential loss. It handles
2017 Dec 14
0
Distributions for gbm models
On page 409 of "Applied Predictive Modeling" by Max Kuhn, it states that the gbm function can accomodate only two class problems when referring to the distribution parameter. >From gbm help re: the distribution parameter: Currently available options are "gaussian" (squared error), "laplace" (absolute loss), "tdist" (t-distribution
2009 Jun 19
2
good boosting tutorial and package in R?
Hi all, Could you please give me some pointers about what's the best boosting package in R currently? in terms of classification accuracy? And any pointers about tutorials and study-materials to curb the learning curve will be greatly appreciated! Thank you! p.s. Does anybody happen to know Boosting implemented in other language such as Matlab? Are they good in terms of accuracy? What
2005 Aug 15
2
How to get a list work in RData file
Dear R-Helper, I want to know how I get a list work which I saved in RData file. For example, > test.xy <- function(x,y) { + xy <- x+y + xy + } > > xyadd <- test.xy(x=2, y=3) > xyadd [1] 5 > x1 <- c(2,43,60,8) > y1 <- c(91,7,5,30) > > xyadd1 <- test.xy(x=x1, y=y1) > xyadd1 [1] 93 50 65 38 > save(list = ls(all=TRUE), file =
2017 Jul 18
1
Help-Multi class classification for large datasets
Hai all, We are working on Multi-class Classification. Currently up to 1.1 million records Ranger package in R is able to handle. Training time on 128 GB RAM is 12 days, which is not a practically feasible method to proceed further. In future we will have dataset of dimension 10 million records, we are in search for a package or framework which can handle 10 million records with at least 12000
2002 Jun 23
2
AdaBoost for R
I'm going to implement AdaBoost algorithm in R. Just wanted to ensure that there is no implementation of any boosting algorithm in R... don't want to reinvent the wheel... -.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.- r-devel mailing list -- Read http://www.ci.tuwien.ac.at/~hornik/R/R-FAQ.html Send "info", "help", or
2005 Jan 25
0
Collapsing solution to the question discussed above: Re: multi-class classification using rpart
You could break your 3 class problem into several (2 or 3) 2 class problems, and then use Andy's suggestion (see the CART book). There are several ways to break the problem into 2 class problems, and several ways to combine the resulting classifiers. Tom Dietterich, Jerry Friedman, Trevor Hastie and Rob Tibshirani, among others, have articles on the question, in places like Annals of
2015 Dec 01
0
[RFC PATCH 5/5] clk: allow boosting only when NvBoost is set
0: disable boosting (cap to base clock from the vbios) 1: boost only to boost clock from the vbios 2: boost to max clock available --- drm/nouveau/include/nvkm/subdev/clk.h | 10 +++++++++- drm/nouveau/nvkm/subdev/clk/base.c | 17 ++++++++++++++++- drm/nouveau/nvkm/subdev/clk/gf100.c | 2 +- drm/nouveau/nvkm/subdev/clk/gk104.c | 2 +- 4 files changed, 27 insertions(+), 4 deletions(-)
2004 Sep 27
1
random discrete from the first tuple
Dear UseR I have a dataset, for instance x1(A, B, C), x2(M,F), x3(X1,X2,X3,X4) and x4(W,F,P). I want to make a small dataset with the random tuple. I know package e1071 can handle a random discrete, > library(e1071) > x1 <- rdiscrete(6, c(2,2,2), c("A","B","C")) > x2 <- rdiscrete(6, c(3,3), c("M","F")) > x3 <-
2015 Dec 02
0
[RFC PATCH 5/5] clk: allow boosting only when NvBoost is set
> Pierre Moreau <pierre.morrow at free.fr> hat am 2. Dezember 2015 um 02:26 > geschrieben: > > Hi Karol, > > I have some comments below. > > On 05:42 PM - Dec 01 2015, Karol Herbst wrote: > > 0: disable boosting (cap to base clock from the vbios) > > 1: boost only to boost clock from the vbios > > 2: boost to max clock available > > --- >