similar to: Problem adding a class to a POSIX vector

Displaying 20 results from an estimated 10000 matches similar to: "Problem adding a class to a POSIX vector"

2003 Apr 21
3
Dates in read.spss
I am using read.spss in the foreign package to read an SPSS save file. For date variables I get huge values such as 11489990400. Does anyone know how to convert these values to R POSIXct date objects? Thanks in advance -Frank platform i686-pc-linux-gnu arch i686 os linux-gnu system i686, linux-gnu status major 1 minor
2001 Apr 07
2
A programming puzzler
I am trying to replicate part of the function of the S-Plus terms.inner function. I don't need an entire model terms object but R expressions or character strings containing the innermost variable name given a vector of character strings or expressions. Here are some example inputs that each should result in the string or expression "x": x; x^2; g(x); h(g(x)); h(g(x^3));
2002 Apr 29
1
Release of Design library; update of Hmisc library
The Design library has been fully ported to R except for Cox proportional hazards regression modeling (using Therneau's survival package) which will be available in about two weeks. It will take much longer to make all the example code executable, is it currently contains many examples for which data are not provided. Thanks to Xiao Gang Fan <xiao.gang.fan1 at libertysurf.fr> who
2002 Apr 29
1
Release of Design library; update of Hmisc library
The Design library has been fully ported to R except for Cox proportional hazards regression modeling (using Therneau's survival package) which will be available in about two weeks. It will take much longer to make all the example code executable, is it currently contains many examples for which data are not provided. Thanks to Xiao Gang Fan <xiao.gang.fan1 at libertysurf.fr> who
2002 Apr 12
5
How to specify search order for require()
In a .First.lib I want to issue two require()s to insure that two other packages are loaded. But I want the package being loaded by .First.lib using library.dynam("mypackage",pkb,lib) to be higher in the search order than the two required packages, because I want to have a couple of functions from the two required packages overridden. What is the best way to do that? Thanks in advance
2002 May 14
2
R CMD check
I am unclear on whether to run R CMD check pgkname as user or as root on Linux. When running as user, after all the latex, html, and man files are created, I get the following error message: Rdconv(): Couldn't open '': Permission denied Has anyone dealt with that message? When I run R CMD check a second time, all latex, html, etc. are recreated which takes quite a while. Is there
2003 Jan 03
4
as.POSIXct problem?
Under platform i686-pc-linux-gnu arch i686 os linux-gnu system i686, linux-gnu status major 1 minor 6.1 year 2002 month 11 day 01 language R > x <- strptime(c('10/10/1969','12/31/2002'),format='%m/%d/%Y')
2002 Mar 28
3
Vectorizing closest match
If anyone has a very fast vectorized method for doing the following I would appreciate some help. I want to avoid outer() to limit memory problems for very large n. Let x = real vector of length n y = real vector of length n w = real vector of length m, m typically less than n/2 but can be > n z = real vector of length m For w[i], i=1,,,m, find the value of x that is closest to w[i]. In
2002 Oct 10
2
tapply for matrices
Does anyone have something like tapply that is extremely fast for matrices when there is a very large number of levels of the grouping variable? I'm referring to, for example, tapply(x, grouping.variable, function.operating.on.submatrix) where x is a matrix and the submatrix is a subset of the rows of x. The grouping variable's length equals the number of rows of x. -- Frank E
2003 Jan 18
1
SAS transport files and the foreign package
Even though the FDA has no policies at all that limit our choices of statistical software, there is one defacto standard in place: reliance of the SAS transport file format for data submission (even though this format is deficient for this purpose, e.g., it does not even document value labels or units of measurement in a self-contained way). Because of the widespread use of SAS transport files in
2002 Dec 04
2
problem with load('http://....') (PR#2344)
Full_Name: Frank Harrell Version: 1.6.1 OS: RedHat 8.0 Linux Submission from: (NULL) (128.143.108.90) I get an error when trying to load a URL that contains a file that was saved using save(object, compress=TRUE): > load('http://hesweb1.med.virginia.edu/biostat/s/data/sav/kprats.sav') Error in gzfile(file, "rb") : unable to open connection In addition: Warning message:
2002 Nov 08
1
pdf() and pdflatex
I am having good success in using pdf() to produce pdf graphics files for inclusion in documents processed by pdflatex [I'm using R 1.6.0 on Linux]. I am having one small problem consistently though, which I believe I also had with S-Plus's pdf.graph(): wasted space at the top of the graph which causes pdflatex to put the graph on a later page, when the current page has plenty of space.
2002 Jul 19
2
Problem with modification of [.factor
Under platform i586-pc-linux-gnu arch i586 os linux-gnu system i586, linux-gnu status major 1 minor 5.0 [1.5.1 not available for Mandrake 8.1] year 2002 month 04 day 29 language R when I modify [.factor to change the default drop=FALSE to
2002 Jul 19
2
Workings of model.frame.default and [.
Related to my earlier question to which I received very helpful replies, when I provide a subsetting method that automatically drops unused levels of a factor variable, I am getting into a bit of trouble using model.frame.default. I know that model.frame.default has its own mechanism for dropping unused levels, but my personal preference is to handle this on a more basic level using [.factor and
2001 Aug 26
1
Re: Variable labels (was Re: [R] Reading SAS version 8 d ata into
> From: fharrell@virginia.edu [mailto:fharrell@virginia.edu] > [snip] > I think your code is more complex that is really needed. > > The problem with defaulting to deparse(...) is that > multiple function pass-throughs return the wrong result: > [snip] > So I don't see a large role for the deparse(...) method. > Actually one of the reasons that I included the
2003 Jun 05
2
ridge regression
Hello R-user I want to compute a multiple regression but I would to include a check for collinearity of the variables. Therefore I would like to use a ridge regression. I tried lm.ridge() but I don't know yet how to get p-values (single Pr() and p of the whole model) out of this model. Can anybody tell me how to get a similar output like the summary(lm(...)) output? Or if there is
2001 Jun 07
1
Suggestion for tapply
In S-Plus the ... argument for tapply is before simplify. In R (1.2.3) ... is after simplify, causing the user to have to specify simplify every time ... is used even though the default is usually OK. Unless there was a benefit to reordering the arguments, it might be a good idea to reconsider the current function definition. ... is probably used more than simplify (e.g., tapply(rainfall,
2003 Jun 17
1
Re: R: Problem from Philippe Glaziou
On Mon, 16 Jun 2003 09:56:07 +0200 Patrick Hausmann <p.hausmann at mr-regionalberatung.de> wrote: > Dear Frank, > > I can reproduce the problem from Philippe Glaziou (see attachment). > Maybe this can help. > Best, > Patrick > > Am Wollelager 11 > 27749 Delmenhorst > Tel. 04221 96373-0 > Fax 04221 96373-29 > > http://www.mr-regionalberatung.de >
2003 Aug 06
3
Slight problem in sort
In platform i686-pc-linux-gnu arch i686 os linux-gnu system i686, linux-gnu status major 1 minor 7.1 year 2003 month 06 day 16 language R I get > sort(c(3,1,NA)) [1] 1 3 Shouldn't NAs be retained by default? Thanks -Frank ---
2002 Aug 04
5
Pseudo R^2 for logit - really naive question
I am using GLM to calculate logit models based on cross-sectional data. I am now down to the hard work of making the results intelligible to very average readers. Is there any way to calculate a psuedo analoque to the R^2 in standard linear regression for use as a purely descriptive statistic of goodness of fit? Most of the readers of my report will be vaguely familiar and more comfortable with