similar to: svm works but tune.svm give error

Displaying 20 results from an estimated 10000 matches similar to: "svm works but tune.svm give error"

2006 Mar 10
1
need help in tune.nnet
Dear R people, I want to use the tune.nnet function of e1071 package to tune nnet . I am unable to understand the parameters of tune.nnet from the e1071 pdf document. I have performed nnet on a traindata and want to test it for class prediction with a testdata. I want to know the values of size,decay,range etc. parameters for which the prediction of testdata is best. Can anyone please tell me
2011 Jan 07
2
Stepwise SVM Variable selection
I have a data set with about 30,000 training cases and 103 variable. I've trained an SVM (using the e1071 package) for a binary classifier {0,1}. The accuracy isn't great. I used a grid search over the C and G parameters with an RBF kernel to find the best settings. I remember that for least squares, R has a nice stepwise function that will try combining subsets of variables to find
2011 May 25
1
help with tune.svm() e1071
Hi, I am trying to use tune.svm in e1071 package. the command i use is tobj <- tune.svm(labels, data= data, cost = 10^(1:2)) Should the last column of the 'data' contain the labels as well? I want to use the linear kernel. But it gives me the error "Error in model.frame.default(formula, data) : 'data' must be a data.frame, not a matrix or an array" Do you know why
2009 Aug 30
1
SVM coefficients
Hello, I'm using the svm function from the e1071 package. It works well and gives me nice results. I'm very curious to see the actual coefficients calculated for each input variable. (Other packages, like RapidMiner, show you this automatically.) I've tried looking at attributes for the model and do see a "coefficients" item, but printing it returns an NULL result.
2009 Aug 04
1
Save model and predictions from svm
Hello, I'm using the e1071 package for training an SVM. It seems to be working well. This question has two parts: 1) Once I've trained an SVM model, I want to USE it within R at a later date to predict various new data. I see the write.svm command, but don't know how to LOAD the model back in so that I can use it tomorrow. How can I do this? 2) I would like to add the
2007 Dec 27
1
(package e1071) SVM tune for best parameters: why they are different everytime i run?
Hi, I run the following tuning function for svm. It's very strange that every time i run this function, the best.parameters give different values. [A] >svm.tune <- tune(svm, train.x, train.y, validation.x=train.x, validation.y=train.y, ranges = list(gamma = 2^(-1:2), cost = 2^(-3:2))) # where train.x and train.y are matrix
2005 May 19
2
tune.svm in {e1071}
Dear All , 1- I'm trying to access the values of fitted(model) after model<- tune.svm( ) but seemingly it is not poosible. How can I access to values of fitted ? However ,it is possible only after model<- svm( ) 2- How can I access to the other values such as the number of Support Vectors , gamma, cost , nu , epsilon , after model<- tune.svm( ) ? these are not possible? I
2009 Aug 12
5
Nominal variables in SVM?
Hi, The answers to my previous question about nominal variables has lead me to a more important question. What is the "best practice" way to feed nominal variable to an SVM. For example: color = ("red, "blue", "green") I could translate that into an index so I wind up with color= (1,2,3) But my concern is that the SVM will now think that the values are
2004 Dec 21
2
Rgui.exe - Error while tuning svm
Hello, if I try to tune my svm with the code: Tune <- tune.svm(Data.Train, Class.Train, type="C-classification", kernel="radial", gamma = 2^(-1:1), cost = 2^(2:4)) i get a windows Messagebox with a error in the application "Rgui.exe" and the message: "Die Anweisung in 0x6c48174d verweist auf Speicher 0x00000000. Der Vorgang "read" konnte nicht auf
2010 Jun 24
1
help in SVM
HI, GUYS, I used the following codes to run SVM and get prediction on new data set hh. dim(all_h) [1] 2034 24 dim(hh) # it contains all the variables besides the variables in all_h data set. [1] 640 415 require(e1071) svm.tune<-tune(svm, as.factor(out) ~ ., data=all_h, ranges=list(gamma=2^(-5:5), cost=2^(-5:5)))# find the best parameters. bestg<-svm.tune$best.parameters[[1]]
2009 Mar 26
1
Extreme AIC in glm(), perfect separation, svm() tuning
Dear List, With regard to the question I previously raised, here is the result I obtained right now, brglm() does help, but there are two situations: 1) Classifiers with extremely high AIC (over 200), no perfect separation, coefficients converge. in this case, using brglm() does help! It stabilize the AIC, and the classification power is better. Code and output: (need to install package:
2009 Nov 17
2
SVM Param Tuning with using SNOW package
Hello, Is the first time I am using SNOW package and I am trying to tune the cost parameter for a linear SVM, where the cost (variable cost1) takes 10 values between 0.5 and 30. I have a large dataset and a pc which is not very powerful, so I need to tune the parameters using both CPUs of the pc. Somehow I cannot manage to do it. It seems that both CPUs are fitting the model for the same values
2010 Jun 17
3
Factoring a variable
Hi, I have a dataset where the results are coded ("yes", "no") We want to do some machine learning with SVM to predict the "yes" outcome My problem is that if I just use the as.factor function to convert, then it reverses the levels. ---------------------- x <- c("no", "no", "no", "yes", "yes", "no",
2004 Dec 16
2
reading svm function in e1071
Hi, If I try to read the codes of functions in e1071 package, it gives me following error message. >library(e1071) > svm function (x, ...) UseMethod("svm") <environment: namespace:e1071> > predict.svm Error: Object "predict.svm" not found > Can someone help me on this how to read the codes of the functions in the e1071 package? Thanks. Raj
2009 Mar 12
0
e1071 SVM one-classification tune problem
Hello all, I am using the e1071 SVM with the tune options for classification, which work pretty well, given the examples of using tune.svm function for classification. But I have not found any example to tune the SVM novelty detection (one-classification) parameters (gamma, cost, nu), for example this are some of the options I have tried with no success: obj<-tune(svm, x,y, type
2006 Dec 08
1
please help me for svm plot question
I run the following code, all other is ok, but plot(m.svm,p5.new,As~Cur) is not ok Anyone know why? install.packages("e1071") library(e1071) library(MASS) p5 <- read.csv("http://www.public.iastate.edu/~aiminy/data/p_5_2.csv") p5.new<-subset(p5,select=-Ms) p5.new$Y<-factor(p5.new$Y) levels(p5.new$Y) <- list(Out=c(1), In=c(0)) attach(p5.new)
2006 Dec 07
1
svm plot question
I run the following code, all other is ok, but plot(m.svm,p5.new,As~Cur) is not ok Anyone know why? install.packages("e1071") library(e1071) library(MASS) p5 <- read.csv("http://www.public.iastate.edu/~aiminy/data/p_5_2.csv") p5.new<-subset(p5,select=-Ms) p5.new$Y<-factor(p5.new$Y) levels(p5.new$Y) <- list(Out=c(1), In=c(0)) attach(p5.new)
2012 Aug 19
1
e1071 - tuning is not giving the best within the range
Hi everybody, I am new in e1071 and with SVMs. I am trying to understand the performance of SVMs but I face with a situation that I thought as not meaningful. I added the R code for you to see what I have done. /set.seed(1234) data <- data.frame( rbind(matrix(rnorm(1500, mean = 10, sd = 5),ncol = 10), matrix(rnorm(1500, mean = 5, sd = 5),ncol = 10))) class <- as.factor(rep(1:2,
2004 Dec 01
1
tuning SVM's
Hi I am doing this sort of thing: POLY: > > obj = best.tune(svm, similarity ~., data = training, kernel = "polynomial") > summary(obj) Call: best.tune(svm, similarity ~ ., data = training, kernel = "polynomial") Parameters: SVM-Type: eps-regression SVM-Kernel: polynomial cost: 1 degree: 3 gamma: 0.04545455 coef.0: 0
2005 Jun 29
2
Running SVM {e1071}
Dear David, Dear Friends, After any running svm I receive different results of Error estimation of 'svm' using 10-fold cross validation. What is the reason ? It is caused by the algorithm, libsvm , e1071 or something els? Which value can be optimal one ? How much run can reach to the optimality.And finally, what is difference between Error estimation of svm using 10-fold cross validation