Displaying 20 results from an estimated 7000 matches similar to: "summary of arima model in R"
2010 Jun 01
1
any doc to understand arima state space model?
I am trying to understand R arima function. Any pointers would be
appreciated.
Thank you,
Shakira.
[[alternative HTML version deleted]]
2004 Nov 15
2
tsdiag() titles
I am using the ts package to fit ARIMA models, and the tsdiag() function to
plot diagnostics. In doing so I'm generating an awful lot of diagnostic
plots of different models and different data sets all within the same R
session. So my question is, is there an option in tsdiag() similar to
<main="Title"> that I can use? This would be quite helpful when I print out
the plots,
2009 Mar 05
2
Overriding contributed package functions
The "tsdiag" function in the TSA package overrides the "tsdiag" function in
the "stats" package. There are a few annoying bugs in the TSA's version of
the function so I would like to use the "stats" function but still have
access to other TSA functions. I have tried using stats::tsdiag( ) but as
long as the TSA package is attached the function from
2012 Mar 29
1
how to increase speed for function?/time efficiency of below function
i am using sarima() function as below
___________________________________________________________________________________________
sarima=function(data,p,d,q,P=0,D=0,Q=0,S=-1,tol=.001){
n=length(data)
constant=1:n
xmean=matrix(1,n,1)
if (d>0 & D>0)
fitit=arima(data, order=c(p,d,q), seasonal=list(order=c(P,D,Q),
period=S),
2003 Dec 08
2
test for arima coef's significancy
Dear sirs,
I would like to know if there is a function to compute the pvalue for the
significancy of arima coef in an arima object created by
the arima function.
I have written this one:
pvalueArima<-function(x,arima)
{
t<-(arima$coef)/(diag(arima$var.coef)^0.5)
df<-length(x)-length(arima$coef)
1-pt(t,df)
}
Has somebody already implemented something equivalent ?
thank you for your
2018 Jan 31
3
Best practices in developing package: From a single file
On 31/01/2018 6:33 AM, Joris Meys wrote:
> 3. given your criticism, I'd like your opinion on where I can improve
> the documentation of https://github.com/CenterForStatistics-UGent/pim.
> I'm currently busy updating the help files for a next release on CRAN,
> so your input is more than welcome.
After this invitation I sent some private comments to Joris. I would
say his
2019 Oct 11
2
New matrix function
I think you are confusing package and function here. Plus some of the R
Core packages, that you mention, contain functions that should probably be
replaced by functions with better implementation from packages on CRAN.
Best regards
Morgan
On Fri, 11 Oct 2019 15:22 Joris Meys, <jorismeys at gmail.com> wrote:
>
>
> On Fri, Oct 11, 2019 at 3:55 PM Morgan Morgan <morgan.emailbox
2013 Feb 28
2
Fortune?
I think the rule is that you can do anything as long as you don't
complain. If you want to complain, you must follow the instructions.
-- Jari Oksanen in
Re: [Rd] Keeping up to date with R-devel
--
Patrick Burns
pburns at pburns.seanet.com
twitter: @burnsstat @portfolioprobe
http://www.portfolioprobe.com/blog
http://www.burns-stat.com
(home of:
'Impatient R'
'The R
2010 May 07
2
help on hmisc
can anyone know where i can find information on compile hmisc on windows, especially 64 windows?
thanks,
_________________________________________________________________
The New Busy is not the too busy. Combine all your e-mail accounts with Hotmail.
ID28326::T:WLMTAGL:ON:WL:en-US:WM_HMP:042010_4
[[alternative HTML version deleted]]
2010 Mar 30
2
weighted.median function from package R.basic
Dear all,
I want to apply a weighted median on a huge dataset, and I remember a
function from the package R.basic that could do this using an internal
sorting algorithm qsort. This speeded things up quite a bit. Alas, I can't
find that package anywhere anymore. There is a weighted.median function in
the package limma too, but I didn't use that before.
Anybody who knows what happened to
2018 Jan 31
3
Best practices in developing package: From a single file
Dear All:
stepping in late, but @Joris, if you would like to take 'from a single file' literally,
have a look at:
https://github.com/bpfaff/lp4rp
(lp4rp: literate programming for R packages);
Cheers,
Bernhard
ps: incidentally, within the noweb-file roxygen is employed.
-----Urspr?ngliche Nachricht-----
Von: R-devel [mailto:r-devel-bounces at r-project.org] Im Auftrag von Joris
2018 Sep 18
3
memory footprint of readRDS()
Dear all,
I tried to read in a 3.8Gb RDS file on a computer with 16Gb available
memory. To my astonishment, the memory footprint of R rises quickly to over
13Gb and the attempt ends with an error that says "cannot allocate vector
of size 5.8Gb".
I would expect that 3 times the memory would be enough to read in that
file, but apparently I was wrong. I checked the memory.limit() and that
2010 May 26
3
cluster analysis and supervised classification: an alternative to knn1?
Hi,
I have a 1.000 observations with 10 attributes (of different types: numeric,
dicotomic, categorical ecc..) and a measure M.
I need to cluster these observations in order to assign a new observation
(with the same 10 attributes but not the measure) to a cluster.
I want to calculate for the new observation a measure as the average of the
meausures M of the observations in the cluster
2009 Feb 03
1
Time series plots with ggplot
Hi,
I am newbie user of ggplot and would like some assistance in
implementing time series plots.
I'd like to know how the tsdiag plot can be made in ggplot?
Thanks
Harsh Singhal
Decisions Systems,
Mu Sigma Inc.
2018 Jun 09
4
Date class shows Inf as NA; this confuses the use of is.na()
And now I've seen I copied the wrong part of ?is.na
> The default method for is.na applied to an atomic vector returns a
logical vector of the same length as its argument x, containing TRUE for
those elements marked NA or, for numeric or complex vectors, NaN, and FALSE
otherwise.
Key point being "atomic vector" here.
On Sat, Jun 9, 2018 at 1:41 PM, Joris Meys <jorismeys at
2010 Jun 03
3
ordinal variables
Dear colleagues,
I teach statistics using SPSS. I want to use R instead. I hit on one problem and I need some quick advice. When I want to work with ordinal variables, in SPSS I can compute the median or create a barchart or compute a spearman correlation with no problems. In R, if I "read" the ordinal variable as numeric, then I cannot do a barplot because I miss the category names. If
2007 Jan 16
2
ARIMA xreg and factors
I am using arima to develop a time series regression model, I am using arima
b/c I have autocorrelated errors. Several of my independent variables are
categorical and I have coded them as factors . When I run ARIMA I don't
get any warning or error message, but I do not seem to get estimates for all
the levels of the factor. Can/how does ARIMA handle factors in xreg?
here is some example
2018 May 02
7
download.file does not process gz files correctly (truncates them?)
Dear all,
I've noticed by trying to download gz files from here :
https://www.ncbi.nlm.nih.gov/geo/query/acc.cgi?acc=GSM907811
At the bottom one can download GSM907811.CEL.gz . If I download this
manually and try
oligo::read.celfiles("GSM907811.CEL.gz")
everything works fine. (oligo is a bioConductor package)
However, if I download using
download.file("
2018 Apr 24
1
data.table not available as win binary for R 3.5 yet?
Dear all,
to my astonishment data.table cannot be installed on R 3.5 Windows. When
checking the package page, the Windows binary is available for download.
When checking the server however, I can't seem to find data.table. Also
install.packages() says the package is only available in source form and
may need compilation.
Compiling using Rtools 35 is no problem.
Is this merely an issue of
2018 Mar 24
1
Possible bug: file.exists() always returns TRUE for prn.us.txt
Thank you. I was just replying my own message with the same information.
Sorry for not doing the research properly before filing.
Cheers
Joris
On Sat, Mar 24, 2018 at 11:36 AM, Duncan Murdoch <murdoch.duncan at gmail.com>
wrote:
> On 24/03/2018 6:16 AM, Joris Meys wrote:
>
>> Dear all,
>>
>> while preparing some exercises I came across some highly surprising