similar to: once more: methods on missing data

Displaying 20 results from an estimated 10000 matches similar to: "once more: methods on missing data"

2001 Jun 06
3
methods on missing data
Hello everybody! I have 2 >issues< concerning methods applied to missing data. I think they're bugs, but who knows. 1. var(NA) returns Error in var(NA) : missing observations in cov/cor instead of NA. I expanded the summary-function to my.summary including SDev, in order to use it with tapply, which crashes in case of groups with no valid data. 2. is a similar problem. I use
2013 Apr 05
1
mixed formatting of integer and numeric (e. g., by summary.default())
Hello, eveRybody, I've been trying to find the origin for the following formatting-"inconsistency": E. g., look at the number of digits in summary.defaults()'s output when NAs occur: in my example below the number of NA's is displayed as an integer, the rest as numeric (floating point numbers): > summary.default( c( 1:2, NA)) Min. 1st Qu. Median Mean 3rd Qu.
2016 Feb 08
1
Apparent bug in summary.data.frame() with columns of Date class and NA's present
Hi all, Based upon an exchange with G?ran Brostr?m on R-Help today: https://stat.ethz.ch/pipermail/r-help/2016-February/435992.html there appears to be a bug in summary.data.frame() in the case where a data frame contains Date class columns that contain NA's and other columns, if present, do not. Example, modified from R-Help: x <- c(18000000, 18810924, 19091227, 19027233, 19310526,
2013 Sep 10
1
[PATCH] show vector length in summary()
(summary.default): show the vector length in addition to quantiles diff -u -i -p -F '^(def' -b -w -B /home/sds/src/R-3.0.1/src/library/base/R/summary.R.old /home/sds/src/R-3.0.1/src/library/base/R/summary.R --- /home/sds/src/R-3.0.1/src/library/base/R/summary.R.old 2013-03-05 18:02:33.000000000 -0500 +++ /home/sds/src/R-3.0.1/src/library/base/R/summary.R 2013-09-10 10:19:02.682946339
2000 Dec 19
1
Bug in glm.fit() or plot.lm() (PR#778)
Here's a bug one of my students noticed. When you call plot() on a glm object, plot.lm gets called. The second plot it shows is supposed to give a normal QQ plot of the standard deviance residuals, but it doesn't. The glm object created by glm.fit returns something (the IRLS weights?) in fit$weights which plot.lm takes as observation weights, so you get strange residuals in the QQ
2005 Apr 28
3
have to point it out again: a distribution question
Stock returns and other financial data have often found to be heavy-tailed. Even Cauchy distributions (without even a first absolute moment) have been entertained as models. Your qq function subtracts numbers on the scale of a normal (0,1) distribution from the input data. When the input data are scaled so that they are insignificant compared to 1, say, then you get essentially the
2010 Aug 24
3
odd behavior of "summary" function
Hello All, Using the standard "summary" function in 'R', I ran across some odd behavior that I cannot understand. Easy to reproduce: Typing: summary(c(6,207936)) Yields:: Min. *1st Qu. Median Mean 3rd Qu. Max.* 6 *51990 104000 104000 156000 207900* None of these values are correct except for the minimum. If I perform "quantile(c(6,
2002 May 17
0
options()$warn==2 and try()
Dear R-help folks: Here is my platform: > version platform sparc-sun-solaris2.7 arch sparc os solaris2.7 system sparc, solaris2.7 status major 1 minor 5.0 year 2002 month 04 day 29 language R I have a
2003 May 29
2
R summary
Dear all i use R only a few days and don't understand the difference between fivenum(x) und summary(x). > x [1] 20.77 22.56 22.71 22.99 26.39 27.08 27.32 27.33 27.57 27.81 28.69 29.36 [13] 30.25 31.89 32.88 33.23 33.28 33.40 33.52 33.83 33.95 34.82 > fivenum(x) [1] 20.770 27.080 29.025 33.280 34.820 > summary(x) Min. 1st Qu. Median Mean 3rd Qu. Max. 20.77 27.14
2008 Nov 20
5
summary statistics into table/data base, many factors to analyse
Dear list, I reduced my data to the following: x <- c(1,4,2,6,8,3,4,2,4,5,1,3) y <- as.factor(c(2,2,1,1,1,2,2,1,1,2,1,2)) z <- as.factor(c(1,2,2,1,1,2,2,3,3,3,3,3)) I can produce the statistical summary just fine. s1 <- tapply(x, y, summary) d1 <- tapply(x, y, sd) s2 <- tapply(x, z, summary) d2 <- tapply(x, z, sd) First thing: I have 100 plus factors to analyse. Theirs
2006 Jul 01
1
noncentral F-distributed random numbers (PR#9055)
Full_Name: Long Qu Version: 2.3.1 OS: Windows XP Submission from: (NULL) (64.113.93.235) The QQ-plot of two versions of simulating noncentral F-distributed random numbers has quite different scales: > qqplot(rf(1000,2,15,3),qf(runif(1000),2,15,3)) The rf() function reads: > rf function (n, df1, df2, ncp = 0) { if (ncp == 0) .Internal(rf(n, df1, df2)) else rchisq(n, df1,
2011 Oct 14
1
Party package: varimp(..., conditional=TRUE) error: term 1 would require 9e+12 columns
I would like to build a forest of regression trees to see how well some covariates predict a response variable and to examine the importance of the covariates. I have a small number of covariates (8) and large number of records (27368). The response and all of the covariates are continuous variables. A cursory examination of the covariates does not suggest they are correlated in a simple fashion
2004 Aug 06
1
solaris 2.7 libshout error
Has anyone seen this issue before on solaris 2.7. I keep getting this error for libshout when trying to compile the example. root#[/usr/local/src/libshout-1.0.9/example]#gcc -lshout -o test example.c Undefined first referenced symbol in file socket /usr/local/lib/gcc-lib/sparc-sun-solaris2.7/3.2/../../../libshout.so recv
2010 Mar 16
2
plm "within" models: is the correct F-statistic reported?
Dear R users I get different F-statistic results for a "within" model, when using "time" or "twoways" effects in plm() [1] and when manually specifying the time control dummies [2]. [1] vignette("plm") [2] http://cran.r-project.org/doc/contrib/Farnsworth-EconometricsInR.pdf Two examples below: library("AER") data("Grunfeld", package =
2003 May 30
0
R summary (and quantiles)
When all else fails, read the help page... ?fivenum says to look at ?boxplot.stats, and the "Details" section of ?boxplot.stats has, well, details. Tukey had reasons to call those hinges rather than quartiles. Andy > -----Original Message----- > From: Knut M. Wittkowski [mailto:kmw at rockefeller.edu] > Sent: Thursday, May 29, 2003 5:58 PM > To: Matthias Kirschner >
2010 May 29
3
adding statistical output to a plot
I have written a function to emulate minitab's QQ plotting output (with SW test and AD test results on the graph): mtab.norm<-function(x) { library(nortest) library(lattice) x<-as.numeric(x) x<-as.vector(x) plot.ht<-4.6 plot.wd<-4.6 pt.ht=plot.ht/5 txt.sz<-(plot.ht/7.5) X11(width=plot.wd, height=plot.ht, bg='gray96') qqplot(x, pch=16, cex=pt.ht,
2007 Aug 03
2
DO NOT REPLY [Bug 4856] New: Filenames are displayed before successful transfer in verbose mode
https://bugzilla.samba.org/show_bug.cgi?id=4856 Summary: Filenames are displayed before successful transfer in verbose mode Product: rsync Version: 2.5.7 Platform: All OS/Version: Linux Status: NEW Severity: normal Priority: P3 Component: core AssignedTo: wayned@samba.org
2009 Jul 09
2
plm Issues
Hi List I'm having difficulty understanding how plm should work with dynamic formulas. See the commands and output below on a standard data set. Notice that the first summary(plm(...)) call returns the same result as the second (it shouldn't if it actually uses the lagged variable requested). The third call results in error (trying to use diff'ed variable in regression) Other info:
2003 May 29
4
Postscript query: plotting long vectors
Hi, I have a query about the maximum length of vector that can be plotted in one go in a postscript driver. Try the following code (in 1.7.0; version details below): t <- seq(from=0, to=4*pi, length=200000) y <- sin(t) postscript(file="o.ps") plot(t, y, type="l") dev.off() If I view the postscript file o.ps in "gv", it takes many seconds before eventually
2011 Jun 30
2
sdev value returned by princomp function (used for PCA)
Dear all, I have a question about the 'sdev' value returned by the princomp function (which does principal components analysis). On the help page for princomp it says 'sdev' is 'the standard deviations of the principal components'. However, when I calculate the principal components for the USArrests data set, I don't find this to be the case: Here is how I