Displaying 20 results from an estimated 120 matches similar to: "Sporadic errors when training models using CARET"
2009 Nov 17
2
SVM Param Tuning with using SNOW package
Hello,
Is the first time I am using SNOW package and I am trying to tune the cost
parameter for a linear SVM, where the cost (variable cost1) takes 10 values
between 0.5 and 30.
I have a large dataset and a pc which is not very powerful, so I need to
tune the parameters using both CPUs of the pc.
Somehow I cannot manage to do it. It seems that both CPUs are fitting the
model for the same values
2010 Nov 24
0
4. Rexcel (Luis Felipe Parra)-how to run a code from excel
Hi Louis,
It's simple to run a r script from the excel spreadsheet.
Just write your code, source("C:\\Quantil
Aplicativos\\Genercauca\\BackwardSelectionNC.r"), into a cell of a
workingsheet. Then right-click the cell and select "run code" in the pop-up
menu.
Hope this will help you.
Best,
Bernard
-----????-----
???: r-help-bounces at r-project.org [mailto:r-help-bounces
2023 May 09
1
RandomForest tuning the parameters
Hi Sacha,
On second thought, perhaps this is more the direction that you want ...
X2 = cbind(X_train,y_train)
colnames(X2)[3] = "y"
regr2<-randomForest(y~x1+x2, data=X2,maxnodes=10, ntree=10)
regr
regr2
#Make prediction
predictions= predict(regr, X_test)
predictions2= predict(regr2, X_test)
HTH,
Eric
On Tue, May 9, 2023 at 6:40?AM Eric Berger <ericjberger at gmail.com>
2012 Nov 23
1
caret train and trainControl
I am used to packages like e1071 where you have a tune step and then pass your tunings to train.
It seems with caret, tuning and training are both handled by train.
I am using train and trainControl to find my hyper parameters like so:
MyTrainControl=trainControl(
method = "cv",
number=5,
returnResamp = "all",
classProbs = TRUE
)
rbfSVM <- train(label~., data =
2009 Jan 15
2
problems with extractPrediction in package caret
Hi list,
I´m working on a predictive modeling task using the caret package.
I found the best model parameters using the train() and trainControl() command. Now I want to evaluate my model and make predictions on a test dataset. I tried to follow the instructions in the manual and the vignettes but unfortunately I´m getting an error message I can`t figure out.
Here is my code:
rfControl <-
2008 Sep 18
1
caret package: arguments passed to the classification or regression routine
Hi,
I am having problems passing arguments to method="gbm" using the train()
function.
I would like to train gbm using the laplace distribution or the quantile
distribution.
here is the code I used and the error:
gbm.test <- train(x.enet, y.matrix[,7],
method="gbm",
distribution=list(name="quantile",alpha=0.5), verbose=FALSE,
2012 Oct 17
0
Superficie de respuesta con rsm y nnet
Hola compañeros de la lista. Los molesto con la siguiente duda.
En un diseño central compuesto (CCD) con dos factores (V1 y V2) y
una variable de respuesta (R), utilizando valores codificados (-1.4142,
-1, 0, 1, 1.4182), al aplicar la orden:
rsm.segundo.orden <- rsm(R ~ Bloque + SO(V1, V2), data =
DATOS.Codificados)
Obtengo el siguiente modelo:
R = 103.92 -2.16
2023 May 08
1
RandomForest tuning the parameters
Dear R-experts,
Here below a toy example with some error messages, especially at the end of the code (Tuning the parameters). Your help to correct my R code would be highly appreciated.
#######################################
#libraries
library(lattice)
library(ggplot2)
library(caret)
library(randomForest)
??
#Data
2017 Nov 24
0
Using bartMachine with the caret package
Dave Langer in this video https://www.youtube.com/watch?v=z8PRU46I3NY
uses the titanic data as an example of using caret to create xgbTree
models. The caret train() function has a tuneGrid parameter which
takes a list set up like so:
tune.grid <- expand.grid(eta = c(0.05, 0.075, 0.1),
nrounds = c(50, 75, 100),
max_depth = 6:8,
2011 May 12
2
Can ROC be used as a metric for optimal model selection for randomForest?
Dear all,
I am using the "caret" Package for predictors selection with a randomForest model. The following is the train function:
rfFit<- train(x=trainRatios, y=trainClass, method="rf", importance = TRUE, do.trace = 100, keep.inbag = TRUE,
tuneGrid = grid, trControl=bootControl, scale = TRUE, metric = "ROC")
I wanted to use ROC as the metric for variable
2010 Mar 23
1
caret package, how can I deal with RFE+SVM wrong message?
Hello,
I am learning caret package, and I want to use the RFE to reduce the
feature. I want to use RFE coupled Random Forest (RFE+FR) to complete this
task. As we know, there are a number of pre-defined sets of functions, like
random Forest(rfFuncs), however,I want to tune the parameters (mtr) when
RFE, and then I write code below, but there is something wrong message, How
can I deal with it?
2013 Mar 06
1
CARET and NNET fail to train a model when the input is high dimensional
The following code fails to train a nnet model in a random dataset using
caret:
nR <- 700
nCol <- 2000
myCtrl <- trainControl(method="cv", number=3, preProcOptions=NULL,
classProbs = TRUE, summaryFunction = twoClassSummary)
trX <- data.frame(replicate(nR, rnorm(nCol)))
trY <- runif(1)*trX[,1]*trX[,2]^2+runif(1)*trX[,3]/trX[,4]
trY <-
2011 Mar 16
1
Specify feature weights in model prediction (CARET)
Using the 'CARET' package, is it possible to specify weights for features
used in model prediction? And for the 'knn' implementation, is there a way
to choose a distance metric (i.e. Mahalanobis distance)?
Thanks,
~Kendric
[[alternative HTML version deleted]]
2010 Oct 22
2
Random Forest AUC
Guys,
I used Random Forest with a couple of data sets I had to predict for binary
response. In all the cases, the AUC of the training set is coming to be 1.
Is this always the case with random forests? Can someone please clarify
this?
I have given a simple example, first using logistic regression and then
using random forests to explain the problem. AUC of the random forest is
coming out to be
2012 May 30
1
caret() train based on cross validation - split dataset to keep sites together?
Hello all,
I have searched and have not yet identified a solution so now I am sending
this message. In short, I need to split my data into training, validation,
and testing subsets that keep all observations from the same sites together
? preferably as part of a cross validation procedure. Now for the longer
version. And I must confess that although my R skills are improving, they
are not so
2012 Feb 10
1
Choosing glmnet lambda values via caret
Usually when using raw glmnet I let the implementation choose the
lambdas. However when training via caret::train the lambda values are
predetermined. Is there any way to have caret defer the lambda
choices to caret::train and thus choose the optimal lambda
dynamically?
--
Yang Zhang
http://yz.mit.edu/
2011 Dec 22
0
randomforest and AUC using 10 fold CV - Plotting results
Here is a snippet to show what i'm trying to do.
library(randomForest)
library(ROCR)
library(caret)
data(iris)
iris <- iris[(iris$Species != "setosa"),]
fit <- randomForest(factor(Species) ~ ., data=iris, ntree=50)
train.predict <- predict(fit,iris,type="prob")[,2]
2010 Sep 29
0
caret package version 4.63
Version 4.63 of the caret package is now on CRAN.
caret can be used to tune the parameters of predictive models using
resampling, estimate variable importance and visualize the results.
There are also various modeling and "helper" functions that can be
useful for training models.
caret has wrappers to over 99 different models for classification
and regression. See the package vignettes
2010 Sep 29
0
caret package version 4.63
Version 4.63 of the caret package is now on CRAN.
caret can be used to tune the parameters of predictive models using
resampling, estimate variable importance and visualize the results.
There are also various modeling and "helper" functions that can be
useful for training models.
caret has wrappers to over 99 different models for classification
and regression. See the package vignettes
2011 Feb 16
1
caret::train() and ctree()
Like earth can be trained simultaneously for degree and nprune, is there a way to train ctree simultaneously for mincriterion and maxdepth?
Also, I notice there are separate methods ctree and ctree2, and if both options are attempted to tune with one method, the summary averages the option it doesn't support. The full log is attached, and notice these lines below for