similar to: need help in tune.nnet

Displaying 20 results from an estimated 3000 matches similar to: "need help in tune.nnet"

2006 Mar 23
0
front- end problem while using nnet and tune.nnet
Dear R people, I am using tune.nnet from e1071 package to tune the parameters for nnet. I am using the following syntax: tuneclass <-c(rep(1,46),rep(2,15)) tunennet <-tune.nnet(x=traindata,y=tuneclass,size=c(50,75,100),decay=c(0,0.005,0.010),MaxNWts = 20000) Here traindata is the training data that I want to tune for nnet which is a matrix with 61 rows(samples) and 200
2006 Feb 02
0
problem with nnet
Hello All, I am working with samr and nnet packages. I am following the steps given below: 1> I take a input file with signal values for 9506 genes and 36 chips , belonging to two classes. 2> I perform samr analysis on 80% of chip data from both the classes.(selected by random sampling) 3> I then use the data of only the significant genes from this samr analysis to train nnet. 4>
2010 Nov 26
1
Issues with nnet.default for regression/classification
Hi, I'm currently trying desperately to get the nnet function for training a neural network (with one hidden layer) to perform a regression task. So I run it like the following: trainednet <- nnet(x=traindata, y=trainresponse, size = 30, linout = TRUE, maxit=1000) (where x is a matrix and y a numerical vector consisting of the target values for one variable) To see whether the network
2006 Jan 05
1
problem with command line arguments
Hello Everybody, I am running a R script through a perl code from command line. The perl script is like: my $cmd= 'R CMD BATCH D:/try5.R'; system($cmd); I run the perl code from command line. Now I want to pass some command line arguments to the R script. Its like the argv concept of perl. Do I pass the arguments through my $cmd in the perl script? If yes, then how to access that in
2013 Mar 06
1
CARET and NNET fail to train a model when the input is high dimensional
The following code fails to train a nnet model in a random dataset using caret: nR <- 700 nCol <- 2000 myCtrl <- trainControl(method="cv", number=3, preProcOptions=NULL, classProbs = TRUE, summaryFunction = twoClassSummary) trX <- data.frame(replicate(nR, rnorm(nCol))) trY <- runif(1)*trX[,1]*trX[,2]^2+runif(1)*trX[,3]/trX[,4] trY <-
2009 Jan 23
1
predict function problem for glmmPQL
Hi all, I am using cross-validation to validate a generalized linear mixed effects model fitted using glmmPQL. i found that the predict function has a problem and i wonder if anyone has encountered the same problem? glmm1 = glmmPQL(y~aX+b,random=~1|sample,data=traindata) predict(glmm1,newdata=testdata,level=1,type="response") gives me all "NA"s. it works for level=0 (the
2020 Jun 05
3
líneas sobre un mapa
Gracias Emilio y Jorge. Tengo que explicarlo mejor. Mostrando a una audiencia cómo hacer un tipo de análisis, se hace un loop (abajo) que analiza un mapa por regiones longitudinales. Tal y como está el script, print(i) te indica la longitud por la que va (de 10º en 10º) pero me gustaría que en vez de eso te fuese representando una línea vertical sobre el mapa, que he representado previamente con
2005 Mar 15
1
KNN one factor predicting problem
Could anybody help me out please? > cl<-as.factor(traindata[,13]) > knn(traindata[1:295,2], newdata[1:32,2], cl,k=2, prob=TRUE) Error in knn(traindata[1:295, 2], newdata[1:32, 2], cl, k = m, prob = TRUE) : Dims of test and train differ Both traindata and newdata have 13 elements. Only one of the first 12 elemnets is needed to predict the 13 element. What's the problem of
2009 Aug 04
1
Strange error with ROCR
Hello, I've come across a strange error... Here is what happens: model <- svm(traindata,trainlabels, type="C-classification", kernel="radial", cost=10, class.weights=c("win"=3,"lose"=1), scale=FALSE, probability = TRUE) predictions <- predict(model, traindata) pred <- prediction(predictions, trainlabels) This returns an error: Error in
2010 Oct 12
1
need help with nnet
HI, Dear R community, My data set has 2409 variables, the last one is response variable. I have used the nnet after feature selection and works. But this time, I am using nnet to fit a model without feature selection. I got the following error information: > dim(train) [1] 1827 2409 nnet.fit<-nnet(as.factor(out) ~ ., data=train, size=3, rang=0.3, decay=5e-4, maxit=500) # model
2009 Aug 25
1
Clogit or LRM?
Hello I believe that I'm getting very close in my modeling application. I've come across a challenge that I am unable to solve and would really appreciate the group's opinion. I've been using the val.prob function from the Design library (Thanks Frank!!) to both evaluate and visualize my model. From the scores and graph, it appears as my model is very accurate in
2008 Sep 25
5
nnet support
Dear Sir/Madam I have recently started using the nnet package but cannot find any documentation other than the one page titled 'nnet {nnet}' which is replicated several times over the internet and is found in the help file for this package. I would like more information on how to use the package and have searched extensively over the internet but cannot find anything more. Do you know of
2003 Oct 20
2
nnet behaving oddly
Hi, I was trying to use the nnet library and am not sure of whats going on. I am calling the nnet function as: n <- nnet(x,y,size=3,subset=sets[[1]], maxit=200) Where x is a 272x4 matrix of observations (examples) and y is a 272x1 matrix of target values. However when I look at nnet$residuals they are off by two orders of magnitude (compared to the output from neural network code that I
2009 Jul 24
1
nnet library and FANN package'm
Hello ! I'd like to know to which of the FANN package network corresponds the R nnet network ? In more details, what is the R nnet activation function, what is the training algorithm (rprop, quickprop, ...) ? Also, it seems that the R nnet "decay" parameter in nnet corresponds to the "learning_rate" parameter in FANN. Correct ? Many thanks in advance ! Luc Moulinier
2008 Feb 15
2
Softmax in nnet
Hi R help, I run my data in nnet with skip layer, factor response (with 0 & 1 values) and explicitly put softmax=T to compare the result of the default nnet with no softmax specification. I assume this should give me the same result. I got the result the default one, but not the softmax version and I got the error message that I did not quite understand. test6.nn.skipT.softm.Yfac <-
2009 Nov 02
1
modifying predict.nnet() to function with errorest()
Greetings, I am having trouble calculating artificial neural network misclassification errors using errorest() from the ipred package. I have had no problems estimating the values with randomForest() or svm(), but can't seem to get it to work with nnet(). I believe this is due to the output of the predict.nnet() function within cv.factor(). Below is a quick example of the problem I'm
2005 Jul 22
2
about nnet package
Dear All, I'm learning to train a neural network with my training data by using nnet package, then evaluate it with a evaluation set. My problem here is that, I need the trained network to be used in future, so, what should I store? and How? Any other options other than nnet package? Any example will be highly appreciated! Best, Baoqiang Cao
2006 Nov 22
1
What training algorithm does nnet package use?
Greetings list, I've just swapped from the "neural" package to the "nnet" package and I've noticed that the training is orders of magnitude faster, and the results are way more accurate. This leads me to wonder, what training algorithm is "nnet" using? Is it a modification on the standard backpropagation? Or a completely different algorithm? I'm
2010 Jun 17
1
help with nnet
> nnet.fit<-nnet(as.factor(out) ~ ., data=all_h, size=5, rang=0.3, decay=5e-4, maxit=500) # model fitting > summary(nnet.fit) a 23-5-1 network with 126 weights options were - entropy fitting decay=5e-04 HI, Guys, I can not find the manual to describe how the model is built, is there a more detailed description how nnet package works? -- Sincerely, Changbin -- [[alternative
2005 Aug 26
1
passing arguments from nnet to optim
Hi everyone, According to R reference manual, the nnet function uses the BFGS method of optim to optimize the neural network parameters. I would like, when calling the function nnet to tell the optim function not to produce the tracing information on the progress of the optimization, or at least to reduce the frequency of the reports. I tried the following: a) nnet default > x<-rnorm(20)