Displaying 20 results from an estimated 20000 matches similar to: "sapply(Date, is.numeric)"
2010 Jun 29
2
POSIXlt matching bug
I came across the below mis-feature/bug using match with POSIXlt objects
(from strptime) in R 2.11.1 (though this appears to be an old issue).
> x <- as.POSIXlt(Sys.Date())
> table <- as.POSIXlt(Sys.Date()+0:5)
> length(x)
[1] 1
> x %in% table # I expect TRUE
[1] FALSE FALSE FALSE FALSE FALSE FALSE FALSE FALSE FALSE
> match(x, table) # I expect 1
[1] NA NA NA NA NA NA NA NA
2010 Jun 23
1
NAs and weighted.mean
R-developers,
In version R 2.11.0, weighted.mean was changed such that:
> weighted.mean(NA, na.rm=TRUE)
[1] 0
rather than NaN as in previous versions of R. I see a note in the NEWS
file indicates that weighted.mean was changed "so an infinite value with
zero weight does not force an NaN result."
In case the side effect of returning 0 rather than NaN in this case was
unintentional,
2008 Aug 07
0
sapply(Date, is.numeric) (PR#12221)
FYI,=20
I've tried posting the below message twice to the bug tracking system,
once by email (below), and the second time 5 days later directly to the
bugs.r-project.org website. As far as I can tell, the bug tracking
system hasn't picked this up. Also it looks like the latest "incoming"
bug is dated 25 May 2008, so perhaps others are having difficulty as
well. (cc: r-bugs)
2008 Aug 07
1
sapply(Date, is.numeric) (PR#12008)
Full_Name: Robert McGehee
Version: 2.7.1
OS: Windows
Submission from: (NULL) (192.223.226.6)
R-developers,
The results below seem inconsistent. From the documentation for is.numeric, I
expect FALSE in both cases.
> x <- data.frame(dt=Sys.Date())
> is.numeric(x$dt)
[1] FALSE
> sapply(x, is.numeric)
dt
TRUE
## Yet, sapply seems aware of the Date class
> sapply(x, class)
dt
2006 Nov 29
1
Removing terms from formula
R-help,
Given a simple linear model, say lm(x ~ y + z), I would like to remove
model terms that are factors with only one level. Thus, if 'z' were a
factor with only one level, lm(x ~ y + z) becomes lm(x ~ y + 1).
Likewise, if both 'y' and 'z' are one-level factors, then the resulting
calculation is simply lm(x ~ 1).
Unfortunately, I have not been able to come up with an
2005 Jan 12
2
?"=" (Windows) (PR#7504)
?"=", ?"==", ?"!=", ?">=", and ?"<=" sends me to the documentation for
?help on Windows, while returning the correct documentation on Linux.
Robert
> version
_
platform i386-pc-mingw32
arch i386
os mingw32
system i386, mingw32
status
major 2
minor
2008 Aug 07
0
sapply(Date, is.numeric) (PR#12239)
>>>>> "RobMcG" == McGehee, Robert <Robert.McGehee at geodecapital.com>
>>>>> on Tue, 29 Jul 2008 15:40:37 -0400 writes:
RobMcG> FYI,
RobMcG> I've tried posting the below message twice to the bug tracking system,
[....... r-bugs problems discussed in a separate thread ....]
RobMcG> R-developers,
RobMcG> The
2008 Aug 07
0
sapply(Date, is.numeric) (PR#12244)
On Wed, 30 Jul 2008, Martin Maechler wrote:
>>>>>> "RobMcG" == McGehee, Robert <Robert.McGehee at geodecapital.com>
>>>>>> on Tue, 29 Jul 2008 15:40:37 -0400 writes:
>
> RobMcG> FYI,
> RobMcG> I've tried posting the below message twice to the bug tracking system,
>
> [....... r-bugs problems discussed in a
2008 Jul 08
1
split.Date
Hello,
I wanted to suggest that the below method for split.Date be added to the
base library to significantly speed up splits with values of class Date.
In the below example I show a speed improvement of 175x for 1000 data
points. On a vector of size 1e6, the time difference was 22 minutes for
split.default versus 0.3 seconds for the split.Date function below (!).
Note that this improvement will
2006 Sep 20
1
seq.Date not accepting NULL length.out (PR#9239)
There seems to be a bug in seq.Date such that it will not allow the user
to pass in length.out =3D NULL, despite the fact that this is the =
default
argument.
For example:
> dt1 <- as.Date("2004-12-31")
> dt2 <- as.Date("2005-12-31")
> seq.Date(dt1, dt2, length.out =3D NULL, by =3D "month")
Error in seq.Date(dt1, dt2, length.out =3D NULL, by =3D
2006 May 22
3
Wishlist: Vignettes on CRAN
I was recently browsing through CRAN's Finance task view to remind
myself of the publicly available packages relevant to my work. As the
reference manuals are all online, I am able to flip through the
available functions to get an idea of the package's scope before
downloading.
That said, many authors have taken the time to additionally provide a
useful vignette which provides a better
2008 Apr 28
2
time zone conversion
Hello,
I'm trying to convert times in the EST/EDT (New York) format to times in
the GMT/BST (London) and UTC+9 format (Tokyo). That is, if I know what
time it is in New York, what is local time in London and Tokyo?
Ex:
Here's the conversion from New York EST/EDT time to London's GMT/BST
time zone for three days in 2007. Note that the US and London change to
daylight savings on
2006 Sep 29
1
Plotting text with lattice
Hello,
I've decided to take the leap and try my hand at the lattice package,
though I am getting stuck at what one might consider a trivial problem,
plotting text at a point in a graph. Apologies in advance if (that) I'm
missing something extremely basic.
Consider in base graphics:
> plot(1:10)
> text(2, 4, "Text")
In the above you will see text centered at the point (2,
2005 Feb 28
2
Changing function arguments to NULL
I'm trying to build a recursive set of functions that take a set of
arguments, change some of the arguments and recursively call the same
(or different) function.
For example here's a stupid recursive counting function that prints back
all integers from x to 0 (and ignores arguments y and z)
cnt <- function(x, y, z) {
stopifnot(is.numeric(x))
print (x)
recursionFUN <-
2008 May 16
1
var/sd and NAs in R2.7.0
Hello all,
I just upgraded to R 2.7.0 and found that the behavior of 'var' and 'sd'
have changed in the presence NAs (this wasn't explicit in the NEWS file,
though I see it probably has to do with the change for cor/cov). Anyway,
I just want to make sure that it was intentional to produce an error
when there was all NAs and na.rm=TRUE, rather than returning an NA (like
R
2005 Jan 27
3
Indexing Lists and Partial Matching
I was unaware until recently that partial matching was used to index
data frames and lists. This is now causing a great deal of problems in
my code as I sometimes index a list without knowing what elements it
contains, expecting a NULL if the column does not exist. However, if
partial matching is used, sometimes R will return an object I do not
want. My question, is there an easy way of getting
2006 Oct 23
4
Changing function arguments
R-Developers,
I'm looking for some help computing on the R language.
I'm hoping to write a function that parses a language or expression
object and returns another expression with all instances of certain
argument of a given function altered. For instance, say I would like my
function, myFun to take an expression and whenever the argument 'x'
appears within the function FUN inside
2005 Apr 22
3
as.data.frame: Error in "names<-.default" (PR#7808)
Hello,
I found a potential problem in R 2.1.0 (and R 2.0.1)
I expect that
> tmp <- FUN(x1, x2, x3, x4)
> as.data.frame(tmp)
is the same as
> as.data.frame(FUN(x1, x2, x3, x4))
since the tmp variable in this case is unnecessary.
However, below I will demonstrate that under an odd set of conditions, I
can correctly perform as.data.frame(tmp), but not as.data.frame(FUN(x1,
x2, x3,
2005 Jan 10
1
new("call") problem (PR#7490)
I have found a solution to the new("call") problem that I believe
produces the correct behavior for the default call object, and am also
reclassifying this as a bug, as I believe the current behavior to be
incorrect.
Recap, the following error occurs:
> new("call")
Error in print("<undef>"()) : couldn't find function "<undef>"
It looks
2008 Sep 12
1
match and incomparables
Hello,
I was playing around with the newly implemented 'incomparables' argument
in 'match' and realized the argument does not behave anything like I
expected. Can someone explain what is going on here? Sorry if I'm
misreading the documentation.
> match(1:3, 1:3, incomparables=1)
[1] NA 2 3 # This seems right, the 1 in 'x' is 'incomparable'
>