similar to: qqnorm() with weighted data?

Displaying 20 results from an estimated 1000 matches similar to: "qqnorm() with weighted data?"

2002 May 01
1
"normal probability plot" with a percentile scale?
I'd like to generate some plots like you'd see on the old "normal probability graph paper", like the first plot in: <http://www.itl.nist.gov/div898/handbook/eda/section3/normprpl.htm> except the horizontal scale would have 1%, 5%, 25%, 50%, 75%, 95%, 99%, or similar quantiles, with associated tick/grid lines. [still hunting around for a good example...] something like
2002 Jul 06
2
cross-platform code: windows or gnu/linux?
I am writing some R code that generates some figures. I am primarily developing on GNU/Linux, but will ultimately run the code on Windows in order to get WMF files I can integrate in an existing document. The win.metafile() function, of course, doesn't exist in the GNU/Linux version of R. I am contemplating writing an abstraction layer that runs the right thing, but I need to know how the
2002 May 02
2
coercing "numeric" components of data frame to "factor" or "ordered"?
I am getting ready to load a bunch of data into R. The data is all numeric, but some of the numbers are integer codes representing non-numeric semantics. What is the best way to "fix" the data frames so that these compenents are recognized as "factors" or "ordered", as appropriate? Can I "assign" to some attribute of the data frame component, like the
2002 May 07
3
good procedure for creating plots for PowerPoint
I am generating some graphs (on GNU/Linux using R 1.5.x) for a PowerPoint presentation my boss is giving next week. He just tried cut/pasted a plot off of a PDF file I had created, but he complained that the fonts were too small and fuzzy, and that he couldn't change the background. I have been playing with png(bg=transparent), but I am not sure what appropriate height,width parameters and
2002 Jun 05
5
hairy indexing problem
I've got a data frame that looks like this: subject foo bar 2 1.7 3.2 2 2.3 4.1 3 7.6 2.3 3 7.1 3.3 3 7.3 2.3 3 7.4 1.3 5 6.2 6.1 5 3.4 6.9 ... That is, I've got multiple rows per subject. I need to compute summaries within categories where the subject has the same number
2002 May 06
3
Spearman rank-order correlation matrix
I"ve got a data frame with a selection of columns I want to compute a rank-order correlation matrix from without disturbing the original data frame. foo[,c("a","b","d","f","g")] What I wanted to do, intuitively, was: > cor(rank(foo[,c("a","b","d","f","g")])) but rank in that context
2002 Jul 09
3
building formula objects
I want to write a function to take an argument as the response variable of a linear model, e.g. to do anova's across a list of variables, something like the following (except, of course, this doesn't work): function(x) { anova(lm(x ~ my.factor,data=my.data)) } The x in lm() above is getting evaluated at the wrong level. How can I make this work? -- Russell Senior ``The two
2002 Apr 04
2
summary on predict with arima0
Here is the summary on predict when using an arima0 object: The arima0 object must be based on a time series vector. That is; x <- ts(xm1, frequency=12, start=c(1975,1)) x.ar <- arima0(x,order=c(1,1,1)) predict(x.ar,n.ahead=3) Thanks so much to Prof. Brian Ripley and David Brahm and other! Sincerely, Erin Hodgess
2003 Oct 28
2
formula parsing, using parts ...
I am writing a little abstraction for a series of tests. For example, I am running an anova and kruskal.test on a one-factor model. That isn't a particular problem, I have an interface like: my.function <- function(model,data) { print(deparse(substitute(data))) a <- anova(lm(formula,data)) print(a) if(a$"Pr(>F)"[1] < 0.05) { pairwise.t.test(???) }
2003 Dec 20
1
why not link devices with --link-dest option?
I notice that devices are not hardlinked when using the --link-dest option, for instance: rsync -aH --link-dest=../bar/ foo/ baz/ When foo/zero is a device, baz/zero is not a hardlink to bar/zero. How come? $ rsync --version rsync version 2.5.6 protocol version 26 on linux 2.4.x -- Russell Senior ``I have nine fingers; you have ten.'' seniorr@aracnet.com
2003 Sep 14
1
estimating quantiles from binned data
Suppose I have a set of binned data, counts exceeding a series of arbitrary thresholds, a total N, a minimum and maximum, those sorts of things. Is there a "standard" method for estimating arbitrary quantiles from this? My initial thought is that the counts and min/max give me solutions at various points along the empirical cdf. As the data are roughly log-normal, I thought maybe I
2007 Feb 21
1
random uniform sample of points on an ellipsoid (e.g. WGS84)
I am interested in making a random sample from a uniform distribution of points over the surface of the earth, using the WGS84 ellipsoid as a model for the earth. I know how to do this for a sphere, but would like to do better. I can supply random numbers, want latitude longitude pairs out. Can anyone point me at a solution? Thanks very much. -- Russell Senior ``I have nine fingers;
2003 Oct 27
2
variance component analysis for nested model
Given a set of data: > names(data) [1] "city" "house" "visit" "value" I am looking for a way to compute the variance components of the nested model (ie, visit 1 at house 2 at city 3 isn't related to visit 1 and house 2 at city 4), but different houses in the same city may be related, and different visits to the same house are probably
2002 Nov 01
2
Empirical distribution
If some can help to find under wich package can i find the commando's wich i schould to use to work with the empirical cummulative (ecdf) distribution and also the QQplot. Zmarrou Hicham Univesity of Amsterdam -.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.- r-help mailing list -- Read http://www.ci.tuwien.ac.at/~hornik/R/R-FAQ.html Send "info",
2003 Oct 22
1
rsync --dry-run --link-dest problem
I am trying to use rsync for making snapshots, sharing hardlinks at the destination using the --link-dest option. In conjunction with this, I would like for --dry-run to report some reliable-ish number for how much space it is going to consume at the destination so that I can arrange to free such space prior the launching rsync "for effect". Right now, that isn't working: #
2004 Jul 03
1
solving for a 2D transformation matrix
We have recently digitized a set of points from some scanned engineering drawings (in the form of PDFs). The digitization resulted in x,y page coordinates for each point. The scans were not aligned perfectly so there is a small rotation, and furthermore each projection (e.g. the yz-plane) on the drawing has a different offset from the page origin to the projection origin. From the dimensions
2010 Apr 15
1
classes and functions for qqnorm and stem
Referring to "Using R for Data Analysis and Graphics" by J H Maindonald, and available from the R site, I found the example on p.30 non-working: > stem(qqnorm(possum$hdlngth)) Error in stem(qqnorm(possum$hdlngth)) : 'x' must be numeric Since qqnorm(possum$hdlngth) plots, and > class(possum$hdlngth) [1] "numeric" , the problem must be here: >
2007 Jun 29
2
why this doesn't work for qqnorm
I want to qqnorm every column in a table. When I try the first column using qqnorm(table$column1), it worked. But when I use qqnorm(table[1]), it tells me "Error in stripchart(x1, ...) : invalid plotting method". What happen? How can I make a function that qqnorms every column? thanks a lot. -jiong The email message (and any attachments) is for the sole use of the intended
2009 Apr 23
1
qqnorm.lme & pairs.lme
Hello, I am trying to do some plotting to check random effect assumptions for a model I fit using lme. I want to use qqnorm and pairs (similarly to examples given in Pinheiro & Bates p. 188), but it's not working. Here's some relevant code and the error message: library(nlme) data(Machines) m1 <- lme(fixed=score~Machine,random=~1|Worker/Machine, data=Machines) qqnorm(m1,
1999 Mar 16
1
qqnorm in R-0.63.3
Dear List, invoking qqnorm-plots in Version 63.3 produces funny things: using the option `type="s"ยด on qqnorm should give a nice *line* of observed quantiles. Now, the line is walking along in order to the points index instead from lowest to highest, wich makes funny slopes. try x <- table(rnorm(1000) # or similar and qqnorm(x,type="s") # in 0.63.2 and 63.3 Well, the