similar to: do_optimhess vs. fdHess ...

Displaying 20 results from an estimated 7000 matches similar to: "do_optimhess vs. fdHess ..."

2013 Jan 22
1
fdHess function
Your question is better addressed to the R-help@R-project.org mailing list, which I am copying on this reply. You are confusing a statistical concept, the Fisher Information matrix, with a numerical concept, the Hessian matrix of a scalar function of a vector argument. The Fisher information matrix is the Hessian matrix of a particular function at its optimum and I have forgotten whether that
2009 Nov 29
1
optim or nlminb for minimization, which to believe?
I have constructed the function mml2 (below) based on the likelihood function described in the minimal latex I have pasted below for anyone who wants to look at it. This function finds parameter estimates for a basic Rasch (IRT) model. Using the function without the gradient, using either nlminb or optim returns the correct parameter estimates and, in the case of optim, the correct standard
2005 Dec 04
1
Understanding nonlinear optimization and Rosenbrock's banana valley function?
GENERAL REFERENCE ON NONLINEAR OPTIMIZATION? What are your favorite references on nonlinear optimization? I like Bates and Watts (1988) Nonlinear Regression Analysis and Its Applications (Wiley), especially for its key insights regarding parameter effects vs. intrinsic curvature. Before I spent time and money on several of the refences cited on the help pages for "optim",
2004 Feb 24
2
convergence in polr
Hello splus-users, I am trying to fit a regression model for an ordered response factor. So I am using the function polr in library(MASS). My data is a matrix of 1665 rows and 63 columns (one of the column is the dependent variable). The code I use is polr(as.ordered(q23p)~.,data=newdatap) but I am getting the following warning message singularity encountered in: nlminb.1(temp, p, liv, lv,
2011 Nov 30
1
How can I pick a matrix from a function? (Out Product of Gradient)
Hi all, I would like to use optim() to estimate the equation by the log-likelihood function and gradient function which I had written. I try to use OPG(Out Product of Gradient) to calculate the Hessian matrix since sometime Hessian matrix is difficult to calculate. Thus I want to pick the Gradient matrix from the gradient function. Moreover, could R show the process of calculation on gradient
2009 Nov 18
1
bug in '...' of constrOptim (PR#14071)
Dear all, There appears to be a bug in how constrOptim handles ... arguments that are suppose to be passed to optim, according to the documentation. This means you can't get the hessian to be returned, for example (so this is a real problem, and not just a question of mistaken documentation). Looking at the code, it appears that a call to the user-defined f includes the ..., when the ...
2007 Sep 16
1
Problem with nlm() function.
In the course of revising a paper I have had occasion to attempt to maximize a rather complicated log likelihood using the function nlm(). This is at the demand of a referee who claims that this will work better than my proposed use of a home- grown implementation of the Levenberg-Marquardt algorithm. I have run into serious hiccups in attempting to apply nlm(). If I provide gradient and
2003 Oct 17
2
nlm, hessian, and derivatives in obj function?
I've been working on a new package and I have a few questions regarding the behaviour of the nlm function. I've been (for better or worse) using the nlm function to fit a linear model without suppling the hessian or gradient attributes in the objective function. I'm curious as to why the nlm requires 31 iterations (for the linear model), and then it doesn't work when I try to add
2004 Sep 27
1
optim error in arima
Hello, I'm fitting a series of ARIMA models to a data set to compare fits. After taking the logs of the data and then differencing them to induce stationarity, I execute arima( y, order=c( p, 0, q ), seasonal=list( order=c( P, 0, Q ), period=7 ) ) for various values of p, q, P and Q. For one set of these values, I get Error in optim(init[mask], armafn, method = "BFGS", hessian
2008 May 23
3
nls diagnostics?
Hi, All: What tools exist for diagnosing singular gradient problems with 'nls'? Consider the following toy example: DF1 <- data.frame(y=1:9, one=rep(1,9)) nlsToyProblem <- nls(y~(a+2*b)*one, DF1, start=list(a=1, b=1), control=nls.control(warnOnly=TRUE)) Error in nlsModel(formula, mf, start, wts) : singular gradient matrix at initial
2003 Dec 02
0
names of parameters from nonlinear model?
I've been trying to figure out how to build a list of terms from a nonlinear model (terms() returns a error). I need to compute and evaluate the partial derivatives (Jacobian) for each equaiton in a set of equations. For example: > eqn <- q ~ s0 + s1 * p + s2 * f + s3 * a > sv2 <- c(d0=3,d1=4.234,d2=4,s0=-2.123,s1=0.234,s2=2.123,s3=4.234) > names( sv2 ) [1] "d0"
2011 Sep 02
5
Hessian Matrix Issue
Dear All, I am running a simulation to obtain coverage probability of Wald type confidence intervals for my parameter d in a function of two parameters (mu,d). I am optimizing it using "optim" method "L-BFGS-B" to obtain MLE. As, I want to invert the Hessian matrix to get Standard errors of the two parameter estimates. However, my Hessian matrix at times becomes
2010 Jun 02
1
Use apply only on non-missing values
I have a function that I am currently using very inefficiently. The following are needed to illustrate the problem: set.seed(12345) dat <- matrix(sample(c(0,1), 110, replace = TRUE), nrow = 11, ncol=10) mis <- sample(1:110, 5) dat[mis] <- NA theta <- rnorm(11) b_vector <- runif(10, -4,4) empty <- which(is.na(t(dat))) So, I have a matrix (dat) with some values within the matrix
2008 Jul 10
0
ace error because of missings?
Hello RUser! I try to use ace for an ancestral state reconstruction but got back an error message. ace(FacVar,Tree, type="discrete") Warning messages: 1: In nlm(function(p) dev(p), p = rep(ip, length.out = np), hessian = TRUE) : NA/Inf durch gr??te positive Zahl ersetzt (NA/Inf replaced by positive number) 2: In nlm(function(p) dev(p), p = rep(ip, length.out = np), hessian
2004 Apr 14
1
How does nlm work?
Dear R users, I have looked in the reference Schnabel, R. B., Koontz, J. E. and Weiss, B. E. (1985) A modular system of algorithms for unconstrained minimization. _ACM Trans. Math. Software_, *11*, 419-440. cited in the nlm help. This article says that the algorithm permits the use of step selection (line search, dogleg and optimal step), analytic or finite diference gradient
2006 Mar 16
1
lme4/Matrix: Call to .Call("mer_update_y"...) and LMEoptimize gives unexpected side effect...
Dear all I want to compute Monte Carlo p-values in lmer-models based on sampled data sets. To speed up calculations, I've tried to use internal functions from the Matrix package (as suggested ealier on the list by Doug Bates). So I did: fm2 <- lmer(resistance ~ ET + position + (1|Grp), Semiconductor,method='ML') simdata<-simulate(fm2,nsim=1) ynew <- simdata[,1] mer
1999 Nov 24
0
nlm gradient and hessian
Out of curiosity, I have tried, without success, to use the new facility in nlm to specify the gradient and hessian. (It is many years since I had a problem simple enough to make analytic derivation of these worthwhile.) The help now says that the function must have attributes with these names but gives no indication as to what should be in the attributes. The online example and demo do not use
2007 Mar 02
2
nlm() problem : extra parameters
Hello: Below is a toy logistic regression problem. When I wrote my own code, Newton-Raphson converged in three iterations using both the gradient and the Hessian and the starting values given below. But I can't get nlm() to work! I would much appreciate any help. > x [1] 10.2 7.7 5.1 3.8 2.6 > y [1] 9 8 3 2 1 > n [1] 10 9 6 8 10 derfs4=function(b,x,y,n) {
2011 May 26
0
Using deriv3() in a separated nonlinear regression model
Hi all, I'm adjusting a nonlinear regression model for data that has a categorigal variable present. So, I can use nls() to do this considering the categorical variable, like this #------------------------------------------------------------ da <- expand.grid(tr=gl(2,1,la=c("tr")), x=1:12) da$y <- 10*da$x/(3+da$x)+rnorm(da$x,0,0.1) plot(y~x, da) n0 <-
2008 Jan 26
1
Any numeric differentiation routine in R for boundary points?
Hi, I have a scalar valued function with several variables. One of the variables is restricted to be non-negative. For example, f(x,y)=sqrt(x)*exp(y), then x should be non-negative. I need the gradient and hessian for some vector (0,y), i.e., I need the gradient and hessian at the boudary of parameter space. The "numderiv" package does not work, even for f(x)=sqrt(x), if you do