Displaying 20 results from an estimated 300 matches similar to: "error with polr()"
2010 Nov 03
2
bugs and misfeatures in polr(MASS).... fixed!
In polr.R the (several) functions gmin and fmin contain the code
> theta <- beta[pc + 1L:q]
> gamm <- c(-100, cumsum(c(theta[1L], exp(theta[-1L]))), 100)
That's bad. There's no reason to suppose beta[pc+1L] is larger than
-100 or that the cumulative sum is smaller than 100. For practical
datasets those assumptions are frequently violated, causing the
2009 Feb 16
0
odd GARCH(1,1) results
Hi everybody,
I'm trying to fit a Garch(1,1) process to the DAX returns. My data
consists of about 2300 10day-logreturns in chronologically descending
order (see attachment). But if I use the garch function I get a very
high alpha_1 and a quite low beta, which doesn't make that much sense. I
think I am missing something, but have no idea what it might be. I'd
appreciate it a lot
2011 Nov 15
2
Controlling the precision of the digits printed
Has anyone come across the right combinations to print a limited number of
digits? My trial and error approach is taking too much time. Here is what I
have tried:
> op <- options()
> a <- c(1e-10,1,2,3,.5,.25)
> names(a) <- c("A", "B", "C", "D", "E", "F")
> # default
> a
A B C D
2009 Apr 22
2
integrate lgamma from 0 to Inf
Dear R users,
i try to integrate lgamma from 0 to Inf. But here i get the message "roundoff error is detected in the extrapolation table", if i use 1.0e120 instead of Inf the computation works, but this is against the suggestion of integrates help information to use Inf explicitly. Using stirlings approximation doesnt bring the solution too.
## Stirlings approximation
lgammaApprox
2010 May 13
1
results of pnorm as either NaN or Inf
I stumbled across this and I am wondering if this is unexpected behavior
or if I am missing something.
> pnorm(-1.0e+307, log.p=TRUE)
[1] -Inf
> pnorm(-1.0e+308, log.p=TRUE)
[1] NaN
Warning message:
In pnorm(q, mean, sd, lower.tail, log.p) : NaNs produced
> pnorm(-1.0e+309, log.p=TRUE)
[1] -Inf
I don't know C and am not that skilled with R, so it would be hard for me
to look into
2009 Jul 27
2
How should i change the SAS Codes into R Codes?
Dear R users,
I have a SAS codes with several loops in it, and i hope to use R to do the
same task. The SAS codes are as follows,
/*to generate the dataset*/
DATA Single_Simulation;
DO se=0 to 1 by 0.01;
DO sp=0 to 1 by 0.01;
DO DR=0 to 1 by 0.01;
TR=(DR+sp-1)/(se+sp-1+1.0e-12);
Adjust_Factor=TR/(DR+1.0e-12);
OUTPUT;
END;
END;
END;
RUN;
/*to select some data*/
DATA
2012 May 08
4
Axes value format
Hi all,
I have some graphs where the values on the X and Y axes are by default
in exponent form like 2e+05 or 1.0e+07. Is it possible to make them in
a more readable form like 10M for 1.0e+07 or 200K for 2e+05?
Thanks and Regards,
- vihan
2012 Jul 02
1
How to get prediction for a variable in WinBUGS?
Dear all,I am a new user of WinBUGS and need your help. After running the following code, I got parameters of beta0 through beta4 (stats, density), but I don't know how to get the prediction of the last value of h, the variable I set to NA and want to model it using the following code.Does anyone can given me a hint? Any advice would be greatly appreciated.Best
2012 Jan 05
2
difference of the multinomial logistic regression results between multinom() function in R and SPSS
Dear all,
I have found some difference of the results between multinom() function in
R and multinomial logistic regression in SPSS software.
The input data, model and parameters are below:
choles <- c(94, 158, 133, 164, 162, 182, 140, 157, 146, 182);
sbp <- c(105, 121, 128, 149, 132, 103, 97, 128, 114, 129);
case <- c(1, 3, 3, 2, 1, 2, 3, 1, 2, 2);
result <- multinom(case ~ choles
2007 Sep 22
1
[LLVMdev] fptoui Semantics Question
I am a little confused by the results of the result of fptoui ... seems to
conflict with the instructions semantics as defined in the language
reference (http://llvm.org/docs/LangRef.html#i_fptoui):
%tmp1001 = fptoui float 1.0E+300 to i1 ; % yields tmp1001 = 0 on my
machine!
but the ref says:
%Y = fptoui float 1.0E+300 to i1 ; yields i1:true
(checked the return value with ... icmp
2011 Jun 26
1
Default values in control list of function
In building a function for a package, I'd like to set the defaults in a control list,
e.g.,
makeg<-function(parameters, eps = 1.0e-7, control=list(showwork=TRUE, rubbish=1.0e+101)){
etc.
}
This does not provide showwork or rubbish within the function if control() is not fully
specified. Copying others, I've previously set the control defaults within the function
then matched
2011 May 27
1
Error with BRugs 0.53 and 0.71, on Win7 with R 2.12.2 and 2.13.0 (crashes R GUI)
I've run into persistent problems with OpenBUGS crashing when using BRugs .53 and .71, and am hoping someone has suggestions. There is obviously something unusual going on in my environment, but I'm at a loss as to where to begin to try to solve it.
In a nutshell, what happens is that, as soon as I call "modelCheck()" in BRugs, it gets an error or crashes ... but only some of
2012 Dec 04
1
Winbugs from R
Hi,
I am trying to covert a Winbugs code into R code. Here is the winbugs code
model{# model’s likelihoodfor (i in 1:n){time[i] ~ dnorm( mu[i], tau ) # stochastic componenent# link and linear predictormu[i] <- beta0 + beta1 * cases[i] + beta2 * distance[i]}# prior distributionstau ~ dgamma( 0.01, 0.01 )beta0 ~ dnorm( 0.0, 1.0E-4)beta1 ~ dnorm( 0.0, 1.0E-4)beta2 ~ dnorm( 0.0, 1.0E-4)#
2011 Mar 28
2
Questions about 'igraph' package.......
I am using 'igraph' package to make some graphs of 'gene-gene interaction'.
I can get a data.frame which has three columns.
gene1 gene2 pvalue
AGT MLR 1.2e-04
MLR 11BHSD1 1.71e-05
IFG2 11BHSD2 2.2e-07
. . .
. . .
. . .
AGTR1 NPPA
2008 Jan 04
2
R2WinBUGS sending variables as factors
Hello R and BUGS users,
I am writing a heirarchical model in R to send to BUGS via R2WinBUGS and
I am finding it difficult to get the model to run. I seem to be having
two problems.
1) I can't seem to send variables classed as factors (Month), is there a
way do this?
2) Checking the Log in WinBUGS I can see that the model is Syntactically
correct, but Bugs is not able to recognise the the
2008 Aug 26
2
Problem with Integrate for NEF-HS distribution
I need to calcuate the cumulative probability for the Natural Exponential Family - Hyperbolic secant distribution with a parameter theta between -pi/2 and pi/2. The integration should be between 0 and 1 as it is a probability.
The function "integrate" works fine when the absolute value of theta is not too large. That is, the NEF-HS distribution is not too skewed. However, once the
2003 Aug 06
3
problems with lda , data included, can somebody test with the new version
enclosed a simple R script (and a data file, and the output) , with
calls lda similar to the example with the iris data in the
documentation. it is not working and i dont understand the error
message. can anybody help me? i am using R 1.5.1 (2002.06.17) on
debian woody stable.
I would like to avoid updating now, because i want to keep the system
in "stable". can somebody test, if the
2019 Jun 30
2
LLVM Releases
I saw an email about 8.0.1 being not quite out, and planning for 9.0.0.
When is 8.0.1 supposed to come out?
I installed llvm at develop on x86 on March 30, and its clang says it is version 9.0.0 (trunk 357348). I tried to do
another spack install llvm at develop, but it said that I already had it installed. Does that mean there
were no changes since March 30, or does spack not look for any
2019 Jul 03
2
LLVM Releases
Thank you.
> On Jul 3, 2019, at 4:04 AM, Justin Clift <justin at postgresql.org> wrote:
>
> On 2019-07-01 00:22, Marty Itzkowitz via llvm-dev wrote:
>> I also tried spack install llvm at develop on a POWER9 (ppc641e)
>> machine, but I can not find a compiler that
>> will compile it. gcc 4.8.5 is reported as too old, and gcc 7.3.0 and
>> 8.1.0 both fail in
2011 Apr 24
2
random roundoff?
On my CentOS 5 box, in a C++ program that does much arithmetic,
including numerous matrix multiplications, I have a situation in
in which the result depends on the nature of nearby I/O. Thus,
with all arithmetic done with type double, and where values
are mostly in the range [-1.0e0,+1.0e0] or nearby, I do:
cerr << "some stuff" << endl;
mat3 = matmult(mat1,mat2);
I