similar to: Error in solve.default(inf, tol = tol.solve) :

Displaying 20 results from an estimated 110 matches similar to: "Error in solve.default(inf, tol = tol.solve) :"

2011 May 04
1
Instrumental variable quantile estimation of spatial autoregressive models
Dear all, I would like to implement a spatial quantile regression using instrumental variable estimation (according to Su and Yang (2007), Instrumental variable quantile estimation of spatial autoregressive models, SMU economics & statistis working paper series, 2007, 05-2007, p.35 ). I am applying the hedonic pricing method on land transactions in Luxembourg. My original data set contains
2020 Feb 17
2
Internal DNS, update of reverse zone fails
On 17/02/2020 10:24, Rowland penny via samba wrote: > On 17/02/2020 08:42, kaffeesurrogat wrote: >> Dear Rowland, >> >> Yes, I did. I'm reading a lot. Docs, books, ... Updates of the >> dns-server via DHCP is up and running, both for the reverse lookup zone >> and the forward lookup mechanism. I've set the lease time to a very low >> value to make
2003 Nov 27
1
lagsarlm - using mixed explanatory variables (spdep package)
Hello I'm very new to R (which is excellent), so apologies if this has already been raised. In the spdep package, I'm trying to undertake an autoregressive mixed model using the lagsarlm function. This is working fine, but there does not appear to be a method of including an explanatory variable without it automatically being included as a lagged term. I'm after something along the
2005 Feb 11
3
How to solve error : "cannot allocate vector of size 1208235 Kb"
Howdy R gurus ! I am newbie to R I use R 2.0.1 in Windows XP. When I run R I got the follwoing memory error. My physical memory size is 3 Gb. My R got the memory problem when it reached to about 2 Gb. Thanks in advance, > library(spdep) > sfr.lagsarlm <- lagsarlm(sfr.data$Bldgsqft ~ sfr.data$Ncounty + sfr.data$Nugb + sfr.data$Ngroup, data=sfr.data, listw=sfr.listw,
2020 Feb 16
3
Internal DNS, update of reverse zone fails
Dear list, one more problem. I've setup my host running a samba addc controller. Samba version is samba-4.11.6-r2. I've joined two win10 clients to my domain. One client has a static ip, the other one was configured to ask my dhcpd-daemon for an ip. Following the book from stefan kania, I modified my dhcpd.conf to execute some scripts I've found on ArchWiki to add my
2010 Dec 09
1
Using Lagsarlm
I'm trying to use the spdep package to calculate this: y = rho W y + e I don't want to use explanatory variables, just the lag from the dependent variable. How would I code this?
2005 Jan 20
0
i want to know about tol in tcp/ip
hello sir.. I am adolf and i am doing my final year engineering in NIT-Trichy. I am going to do my project in job scheduling. Fot that i want to know that how to extract the tol from the tcp/ip and some more details of that. It will be very helpful for my project. Bye. ADOLF. _______________________________________________ LARTC mailing list / LARTC@mailman.ds9a.nl
2016 Mar 24
0
summary( prcomp(*, tol = .) ) -- and 'rank.'
Martin, I fully agree. This becomes an issue when you have big matrices. (Note that there are awesome methods for actually only computing a small number of PCs (unlike your code which uses svn which gets all of them); these are available in various CRAN packages). Best, Kasper On Thu, Mar 24, 2016 at 1:09 PM, Martin Maechler <maechler at stat.math.ethz.ch > wrote: > Following from
2012 Nov 16
1
tol in prcomp
Hi there, I was wondering if anyone could explain how you should set tol in the prcomp function. Using help(prcomp) explains it as "a value indicating the magnitude below which components should be omitted. (Components are omitted if their standard deviations are less than or equal to tol times the standard deviation of the first component.) With the default null setting, no components
2007 Jan 03
1
Visual Studio project files in Speex 1.2beta
Hello, I maintain audio processing part of code in sipXtapi -- open-source VoIP library. And when I engaged Speex library, I found that its Windows project files are not up to date and use strange settings. They miss several source files and use single-threaded static runtime libraries. I see no benefits of using static and moreover single-threaded libraries. In the modern world very small amount
2016 Mar 25
0
summary( prcomp(*, tol = .) ) -- and 'rank.'
As I see it, the display showing the first p << n PCs adding up to 100% of the variance is plainly wrong. I suspect it comes about via a mental short-circuit: If we try to control p using a tolerance, then that amounts to saying that the remaining PCs are effectively zero-variance, but that is (usually) not the intention at all. The common case is that the remainder terms have a roughly
2007 May 29
3
Speex.lib
Hi, Thanks! can this program generate speex.lib instead of libspeex.lib?? the OpenSpeak link to speex.lib instead of libspeex.lib I can compiled the libspeex.lib but not speex.lib. Correct me if I am wrong Thanks ----- Original Message ----- From: "Alexander Chemeris" <ipse.sipx@gmail.com> To: <alex@gorex.com.hk> Cc: <speex-dev@xiph.org> Sent: Tuesday, May 29,
2016 Mar 25
0
summary( prcomp(*, tol = .) ) -- and 'rank.'
> On 25 Mar 2016, at 10:08 , Jari Oksanen <jari.oksanen at oulu.fi> wrote: > >> >> On 25 Mar 2016, at 10:41 am, peter dalgaard <pdalgd at gmail.com> wrote: >> >> As I see it, the display showing the first p << n PCs adding up to 100% of the variance is plainly wrong. >> >> I suspect it comes about via a mental short-circuit: If we
2020 Feb 17
0
Internal DNS, update of reverse zone fails
On 17/02/2020 10:01, kaffeesurrogat via samba wrote: > > On 17/02/2020 10:24, Rowland penny via samba wrote: >> On 17/02/2020 08:42, kaffeesurrogat wrote: >>> Dear Rowland, >>> >>> Yes, I did. I'm reading a lot. Docs, books, ... Updates of the >>> dns-server via DHCP is up and running, both for the reverse lookup zone >>> and the forward
2016 Mar 24
3
summary( prcomp(*, tol = .) ) -- and 'rank.'
Following from the R-help thread of March 22 on "Memory usage in prcomp", I've started looking into adding an optional 'rank.' argument to prcomp allowing to more efficiently get only a few PCs instead of the full p PCs, say when p = 1000 and you know you only want 5 PCs. (https://stat.ethz.ch/pipermail/r-help/2016-March/437228.html As it was mentioned, we already
2004 Nov 18
1
Method dispatch S3/S4 through optimize()
I have been running into difficulties with dispatching on an S4 class defined in the SparseM package, when the method calls are inside a function passed as the f= argument to optimize() in functions in the spdep package. The S4 methods are typically defined as: setMethod("det","matrix.csr", function(x, ...) det(chol(x))^2) that is within setMethod() rather than by name before
2016 Mar 25
2
summary( prcomp(*, tol = .) ) -- and 'rank.'
> On 25 Mar 2016, at 10:41 am, peter dalgaard <pdalgd at gmail.com> wrote: > > As I see it, the display showing the first p << n PCs adding up to 100% of the variance is plainly wrong. > > I suspect it comes about via a mental short-circuit: If we try to control p using a tolerance, then that amounts to saying that the remaining PCs are effectively zero-variance, but
2016 Mar 24
3
summary( prcomp(*, tol = .) ) -- and 'rank.'
I agree with Kasper, this is a 'big' issue. Does your method of taking only n PCs reduce the load on memory? The new addition to the summary looks like a good idea, but Proportion of Variance as you describe it may be confusing to new users. Am I correct in saying Proportion of variance describes the amount of variance with respect to the number of components the user chooses to show? So
2012 Jul 24
1
unable to run spatial lag and error models on large data
Hi: First my apologies for cross-posting. A few days back I posted my queries ar R-sig-geo but did not get any response. Hence this post. I am working on two parcel-level housing dataset to estimate the impact of various variables on home sale prices. I created the spatial weight metrics in ArcGIS 10 using sale year of four nearest houses to assign weights.  Next, I ran LM tests and then ran
2006 Jul 13
0
SAR with count data
I would like to undertake a Spatial autoregression with count data, how can I specify a negative binomial distribution? I am using the lagsarlm function in the spdep package. Thanks Julie -- View this message in context: http://www.nabble.com/SAR-with-count-data-tf1936174.html#a5304944 Sent from the R help forum at Nabble.com.