similar to: Significance of confidence intervals in the Non-Linear Least Squares Program.

Displaying 20 results from an estimated 9000 matches similar to: "Significance of confidence intervals in the Non-Linear Least Squares Program."

2008 Mar 27
1
[Re: Significance of confidence intervals in the Non-Linear Least Squares Program.]
Thanks for the response. I was not very clear in my original request. What I am asking is if in a non-linear estimation problem using nls(), as the condition number of the Hessian matrix becomes larger, will the t-values of one or more of the parameters being estimated in general become smaller in absolute value -- that is, are low t-values a sign of an ill-conditioned Hessian? Typical
2020 Oct 15
2
package(moments) issue
Hi all, While running the anscombe.test in R, I'm getting an error of *Error in if (pval > 1) pval <- 2 - pval : missing value where TRUE/FALSE needed* for a few time series columns whereas for most of the series the function is working fine. I have checked for those specific columns for missing values. However, there is no NA/NAN value in the dataset. I have also run kurtosis for
2012 Apr 19
3
Solve an ordinary or generalized eigenvalue problem in R?
Folks: I'm trying to port some code from python over to R, and I'm running into a wall finding R code that can solve a generalized eigenvalue problem following this function model: http://docs.scipy.org/doc/scipy/reference/generated/scipy.linalg.eig.html Any ideas? I don't want to call python from within R for various reasons, I'd prefer a "native" R solution if one
2011 May 04
1
bivariate linear interpolation
Hi, I have three matrices (X,Y,P) with the same dimension. The X,Y grid is regular and I want to perform linear interpolation to pick out certain points. In matlab appropriate call is something like Pout=interp2(X,Y,P,Xout,Yout, method="linear") where Xout and Yout are the locations where I want the Pout data (typically a different grid). (Scipy has this routine in
2023 Mar 01
1
Incorrect behavior of ks.test and psmirnov functions with exact=TRUE
HI, I've noticed what I think is an incorrect behavior of stats::psmirnov function and consequently of ks.test when run in an exact mode. For example: psmirnov(1, sizes=c(50, 50), z=1:100, two.sided = FALSE, lower.tail = F, exact=TRUE) produces 2.775558e-15 However, the exact value should be 1/combination(100, 50), which is 9.9e-30. While the absolute error is small, the relative error is
2020 Oct 15
0
package(moments) issue
moments::anscombe.test(x) does give errors when x has too few values or if all the values in x are the same > moments::anscombe.test(c(1,2,6)) Error in if (pval > 1) pval <- 2 - pval : missing value where TRUE/FALSE needed > moments::anscombe.test(c(2,2,2,2,2,2,2,2)) Error in if (pval > 1) pval <- 2 - pval : missing value where TRUE/FALSE needed You can use tryCatch() to
2020 Oct 15
2
package(moments) issue
Hi Bill, Thanks for prompt reply and letting me know a way around it. I have more than 1200 observations and not all the values are the same. However, my data points are quite similar, for example, 0.079275, 0.078867, 0.070716 in millions and etc. I have run the data without converting it to millions and I still get the same error message. As I have kurtosis value, it should be fine for the
2012 Feb 04
1
least squares solution to linear system
Dear all I am having a linear system of the form A*X=B and I want to find the X by using least squares. For example my A is of dimension [205,3] and my B is of dimension[205,1]  I am looking for the X matrix which is of the size [3,1]. In the matlab I was doing that by the function  X = LSCOV(A,B) returns the ordinary least squares solution to the     linear system of equations A*X = B, i.e., X
2012 Oct 19
2
Which package/function for solving weighted linear least squares with inequality and equality constraints?
Dear All, Which package/function could i use to solve following linear least square problem? A over determined system of linear equations is given. The nnls-function may would be a possibility BUT: The solving is constrained with a inequality that all unknowns are >= 0 and a equality that the sum of all unknowns is 1 The influence of the equations according to the solving process is
2006 Feb 21
2
How to get around heteroscedasticity with non-linear least squares in R?
I am using "nls" to fit dose-response curves but am not sure how to approach more robust regression in R to get around the problem of the my error showing increased variance with increasing dose. My understanding is that "rlm" or "lqs" would not be a good idea here. 'Fairly new to regression work, so apologies if I'm missing something obvious.
2010 Jan 28
3
weighted least squares vs linear regression
I need to find out the difference between the way R calculates weighted regression and standard regression. I want to plot a 95% confidence interval around an estimte i got from least squares regression. I cant find he documentation for this ive looked in ?stats ?lm ?predict.lm ?weights ?residuals.lm Can anyone shed light? thanks Chris. -- View this message in context:
2005 Mar 17
1
Optimization of constrained linear least-squares problem
Dear R-ians, I want to perform an linear unmixing of image pixels in fractions of pure endmembers. Therefore I need to perform a constrained linear least-squares problem that looks like : min || Cx - d || ? where sum(x) = 1. I have a 3x3 matrix C, containing the values for endmembers and I have a 3x1 column vector d (for every pixel in the image). In theory my x values should all be in the
2009 May 07
2
Linear least squares fit with errors in both x and y values.
HI, I'd like to perform a weighted linear least squares fit with R on data with varying errors on both vectors. I can do this with one axis using lm, but have no idea where to go from here. I've tried googling, but no idea. Any suggestions? Thanks, James
2007 Sep 05
2
question about non-linear least squares in R
Hi, everyone, My question is: It's not every time that you can get a converged result from the nls function. Is there any solution for me to get a reasonable result? For example: x <- c(-0.06,-0.04,-0.025,-0.015,-0.005,0.005,0.015,0.025,0.04,0.06) y <- c(1866760,1457870,1314960,1250560,1184850,1144920,1158850,1199910,1263850,1452520) fitOup<- nls(y ~ constant + A*(x-MA)^4 +
2020 Oct 15
0
package(moments) issue
Another bad case is > moments::anscombe.test(rep(c(1,1.1),length=35)) Error in if (pval > 1) pval <- 2 - pval : missing value where TRUE/FALSE needed I haven't checked the formulas carefully, but I suspect the problem is from taking the cube root of a negative number in z <- (1 - 2/(9 * a) - ((1 - 2/a)/(1 + xx * sqrt(2/(a - 4))))^(1/3))/sqrt(2/(9 * a)) In R, the
2007 Nov 28
2
alternatives to traditional least squares method in linear regression ?
Dear list, I have encountered a special case for searching a linear regression where I'm not satisfied with the results obtained using the traditional least squares method (sometimes called OLS) for estimating/optimizing the residues to the regression line (see code below). Basically, a group of my x-y data are a bit off the diagonal line (in my case the diagonal represents the ideal or
2019 Jun 23
2
Calculation of e^{z^2/2} for a normal deviate z
I agree with many the sentiments about the wisdom of computing very small p-values (although the example below may win some kind of a prize: I've seen people talking about p-values of the order of 10^(-2000), but never 10^(-(10^8)) !). That said, there are a several tricks for getting more reasonable sums of very small probabilities. The first is to scale the p-values by dividing the
2020 Jun 23
8
[Incubation] Request to incubate mlir-npcomp
Per the recent (seeming) consensus regarding incubating new projects under the LLVM organization, I would like to trial the process by requesting to incubate mlir-npcomp <https://github.com/google/mlir-npcomp>. The project is still quite young and has been primarily developed part time by myself and Sean Silva over the last ~2 months. We set it up following discussion of a Numpy/Scipy op set
2010 Feb 10
2
Total least squares linear regression
Dear all, After a thorough research, I still find myself unable to find a function that does linear regression of 2 vectors of data using the "total least squares", also called "orthogonal regression" (see : http://en.wikipedia.org/wiki/Total_least_squares) instead of the "ordinary least squares" method. Indeed, the "lm" function has a
2009 Nov 21
3
python
Dear R users, I would like to make my R code for MCMC faster. It is possible to integrate C code into R but I think C is too complicated for me. I would need a C introduction only for MCMC and I do not know if such a thing exists. I was thinking of Python (and scipy). Where could I read about its integration into R ? How developed are the statistical packages in Python ? I could not find a