similar to: frequency table

Displaying 20 results from an estimated 2000 matches similar to: "frequency table"

2002 Oct 14
2
Another newbie question: curve of normal distribution
I would like to get a curve of normal distrubtion over the histogram. Something like the following (which obviously doesn't work; see attached example). maluj <- function() { vrhy=read.csv("pennies.csv",head=TRUE) hf=table(vrhy$HEADS) postscript("heads.eps",onefile=FALSE,width=4.134,height=3.445,pointsize=12) plot(hf,main="Frequency distribution of
2002 Oct 23
7
Counting NA?
Hi, how to do quickly equivalent of the following? counter = 0 for(i in 1:length(data$S2)) { if(!is.na(data$S2[i])) { counter = counter + 1 } } I have imagined something like length(x,na.rm=TRUE). How can I get values usually taken from tables like z-score, values of t distribution etc.? I could not find them among values mentioned in info file. Thanks, Matej -- Matej Cepl,
2002 Nov 13
4
[Newbie] Is there any support for work with grouped frequencies?
Hi, I have this table (BTW, published by FBI and representing age distribution of the WTC tragedy victims): "LABEL" "FREQ" "MIDPOINT" "1" "Infant (under 1)" 0 "2" "1 to 4" 5 "3" "5 to 8" 1 "4" "9 to 12" 3 "5" "13 to 16" 0 "6" "17 to
2002 Nov 15
2
What's going on? (strange standard deviation)
I have probably misunderstood something (again :-(): > CEMPLOY<-TEMPLOY[!is.na(TEMPLOY)] > sum(CEMPLOY^2/length(CEMPLOY))-mean(CEMPLOY)^2 [1] 30877.28 > var(TEMPLOY,na.rm=TRUE) [1] 30981.25 Could anybody tell me, what's going on? I suppose that var=sum(x^2/N)-xbar^2 or what did I enter wrong? Thanks Matej -- Matej Cepl, matej at ceplovi.cz, Finger: 89EF 4BC6 288A
2002 Nov 21
1
Pearson's correlation coefficient?
How do I get non-squared correlation coefficient in some more sensible way than sqrt(summary(lm(y~x))$r.squared)? Thanks Matej -- Matej Cepl, matej at ceplovi.cz, Finger: 89EF 4BC6 288A BF43 1BAB 25C3 E09F EF25 D964 84AC 138 Highland Ave. #10, Somerville, Ma 02143, (617) 623-1488 The difference between death and taxes is death doesn't get worse every time Congress meets -- Will
2002 Nov 26
4
[NB] lm problems
Hi, I have probably overlooked something obvious, but could anybody help me with following, please? Trying to make regression analysis. I have a huge dataframe with results from National Opinion Survey on Crime and Justice (www.abacon.com/fox/) with two variables G5 and N3 which are imported to R as ordered factors: > levels(noscj$G5) [1] "Strongly agree" "Agree"
2004 Apr 22
1
Selection of cities sample
Hi, I have a question, how to most properly select set of cities which would be as similar as possible in some particular variables with the City of Boston (which I use as my base line). I thought about ordering cities by sum of ((differences between value of that particular variable for that particular city and the value of same variable for Boston) divided by the standard deviation of the
2006 Aug 04
3
Lyrics stored in .ogg files?
Hi, I've heard somewhere that it is possible to store inside .ogg files text streams and thus it should be possible to store lyrics for the sond directly in the file. Is it true? Could anybody point me towards some code how to do it (Python preferable, but not required)? Is there any documentation for text streams in .ogg files? Thanks for any reply, Mat?j -- GPG Finger: 89EF 4BC6 288A
2002 Dec 02
2
Crawley's book on S-Plus and one strangeness
Hi, I have got to my hands an excellent book by Michael J. Crawley ``Statistical Computing: An Introduction to Data Analysis using S-Plus'' (John Wiley & Sons, Ltd, ISBN 0-471-56040-5). Its beauty for me is in the fact, that it is more of ``An Introduction to Data Analysis'' than ``using S-Plus'', but I guess that it may be of interest for many others. Most of the
2010 Jun 08
8
[Bug 28449] New: I messed up and now nouveau doesn't work.
https://bugs.freedesktop.org/show_bug.cgi?id=28449 Summary: I messed up and now nouveau doesn't work. Product: xorg Version: git Platform: x86 (IA32) OS/Version: Linux (All) Status: NEW Severity: critical Priority: medium Component: Driver/nouveau AssignedTo: nouveau at lists.freedesktop.org
2002 Nov 19
5
plotting intersecting planes
Hi all In two days, I am giving a small, informal workshop in our Department about using R. Initially, it was just for statistician, but surprisingly (to me anyhow) many mathematicians are also coming who have a MATLAB background. They are coming at the Workshop from a teaching perspective. They are considering using R to avoid licensing issues with MATLAB. One thing they were hoping me to
2002 Nov 23
1
t-test for correlation matrix?
Hi, I understand that I can find t-test for the significance of the correlation coefficient on sample to population, but is there way how to make this test for more than one pair of variables together as easily as I can make a correlation matrix for more variables with cor(cbind(A,B,C,D))? Thanks, Matej -- Matej Cepl, matej at ceplovi.cz, Finger: 89EF 4BC6 288A BF43 1BAB 25C3 E09F EF25 D964
2002 Nov 06
2
Re: some questions!
Hi, I'm also cc'ing it to r-help. On Wed, 6 Nov 2002, Kenneth Cabrera wrote: > Date: Wed, 06 Nov 2002 10:09:21 -0500 > From: Kenneth Cabrera <krcabrer at perseus.unalmed.edu.co> > To: kwan022 at stat.auckland.ac.nz > Subject: some questions! > > Hello Dear Ko-Kang Wang: > > I am trying to compile R v 1.6.1 . > > I am following the
2002 Oct 09
1
Multiple plots
Hi, I would love to make multiple histograms transposed one on another in order to show relation between the sets. I tried to write a function like this, but R tells me, that I cannot use add=FALSE in high-level commands. That's nice but I am supposed to do? rm(list=ls()) # what's wrong with underscore? #getwd("/home/matej/docs/skola/stat\_anal-cj3534/assign01/")
2002 Oct 28
3
changing coordinates?
I just detected R and have, after browsing the manual, one question: I look for quite a lot of time for graphical software which allows to plot data from a table or external file _with the axes of the coordinate system changed_, i.e. the x-axis should run from top left to bottom left, and the y-axis from top-left to top-right. It is of no use to interchange the rows in the table, the
2002 Oct 26
2
Still missing something on missing values...
Hi, I have a SPSS datafile which is used for my textbook in the statistics (and which is available on http://abacon.com/fox/s6720p2.sav, but it is originally from ICPSR). When I opened it with SPSS 10 and run Frequencies on it I have got 979 valid data a 27 missing. However, see below (unfortunately, I have used R in preparation of my homework, which caused me an error on this): >
2004 Apr 27
2
minimal requirement
Hi, what is the minimal hardware requirement for run R on windows 9x? and for run it in a linux with X and a light windowmanager? Not for hard use, only for learning. Thanks Ronaldo -- Genius may have its limitations, but stupidity is not thus handicapped. -- Elbert Hubbard -- |> // | \\ [***********************************] | ( ?? ?? ) [Ronaldo Reis J??nior ]
2004 May 21
1
No load() from script
Hi, when I try to load data from a file _in a script_ (run either with R --vanilla < script.R, or R BATCH --vanilla script.R, or even with source() from inside R), then I got this error message and scripts halts > #data.multiple <- read.table(file="multiple.csv",sep="\t") > load(file="multiple.RData") Error in open.connection(con, "rb") :
2020 Nov 20
2
return (x+1) * 1000
Without having dug into the details, it could be that one could update the parser by making a 'return' a keyword and require it to be followed by a parenthesis that optionally contains an expression followed by end of statement (newline or semicolon). Such a "promotion" of the 'return' statement seems backward compatible and would end up throwing syntax errors on:
2020 Nov 21
1
return (x+1) * 1000
On 20/11/2020 7:01 p.m., Ben Bolker wrote: > I may be unusual but I don't find these examples surprising at all/ > I don't think I would make these mistakes (maybe it's easier to make > that mistake if you're used to a language where 'return' is a keyword > rather than a function? > > My two cents would be that it would make more sense to (1) write