Displaying 20 results from an estimated 2000 matches similar to: "Adding summary title to table"
2012 Sep 03
3
Horizontal grid in background of barplot
All,
I have:
x <- matrix(c(22,3,6,69,9,4,7,81,23,4,22,50),nrow=3,byrow=TRUE)
rownames(x) <- c("Cold or flu","Headache","Backache");
colnames(x) <- c("Went to doctor","No response","Did
nothing","Self-medicated")
x <- t(x)
print(x)
barplot(x,beside=TRUE,
ylim=c(0,90),
xlab="Ailment",
2009 Oct 14
1
Beginning R help?
Hi all,
I'm currently working through the "Beginner's Guide to R" (Zurr et al.) and
I'm wondering about the first exercise in chapter 3:
I imported the data from BirdFluCases.txt and executed the 'names' and 'str'
functions as follows:
Bird = read.table(file='C:\\rbook\\BirdFluCases2.txt', header=TRUE)
names(Bird)
[1] "Year"
2009 Oct 14
1
R neophyte question.
Hi all,
I'm currently working through the "Beginner's Guide to R" (Zurr et al.) and
I'm wondering about the first exercise in chapter 3:
I imported the data from BirdFluCases.txt and executed the 'names' and 'str'
functions as follows:
Bird = read.table(file='C:\\rbook\\BirdFluCases2.txt', header=TRUE)
names(Bird)
[1] "Year"
2007 Aug 02
1
New user help with plot.default
Hi - I'm just starting out with R and have come up against a problem with
what should be a simple operation. I'm plotting a range of safety standards
for 7 different chemicals. I don't want the standard box and whisker plot -
I just a dot for each value. Whilst using "plot.default" I get the error
"invalid xlim value" - which is confusing since using the same script
2014 Feb 04
2
Status
So, just some quick notes:
- dovecot.org server was down a couple of days again, need to find time to move it to a VM..
- my tss at iki.fi emails were also broken for some hours at least and it was bouncing back all mails to it during the time
- I fixed some bugs reported by people, but I did it in airplane so couldn't reply back with links to the fixes in hg repo. I'll probably do
2010 Apr 05
1
back soon / clang static analyzer
I had a flu (or something) last week and couldn't get almost anything
done. I'm pretty much healed now, and today I spent some time fixing
bugs already, but then I got stuck fixing issues reported by clang's
static analyzer. The good thing is that it found several real bugs from
error handling paths too.
I'm now planning on getting the false positive list down to zero and
maybe
2004 Jun 29
1
strucchange-esque inference for glms ?
hello R-world,
according to the strucchange package .pdf, "all procedures in this package are
concerned with testing or assessing deviations from stability in the classical
linear regression model."
i'd like to test/assess deviations from stability in the Poisson model.
is there a way to modify the strucchange package to suit my purposes, or should
i use be using another
2018 May 09
3
Seasonal weekly average
Hi,
I am fairly new to 'R' and would like advice on the following. I want to calculate a weekly average number of reports (e.g. of flu, norovirus) based on the same weeks for the last five years. I will then use this to plot a chart with 52 points for the average based on the last five years; another line will then plot the current year, enabling a comparison of current weekly counts
2005 Mar 07
3
Tons of shiny new stuff
Ok despite this crazy flu which got me last week i was able to get
some coding done.
Here goes:
== Hieraki
Hieraki got plenty of improvements next to the rails 0.10 update. The
biggest new features are proper yaml, html and readme export. But
there are also plenty of new code improvements like the switch to my
favorite postback style for all controllers. Hieraki traditionally
serves as a demo
2017 Oct 21
1
Skip error in downloading file in loop
I am trying to download data from NASA web-service.
I am using the following code;
for( i in 1:8) {
target1 <- paste0("
https://hydro1.gesdisc.eosdis.nasa.gov/daac-bin/access/timeseries.cgi?variable=NLDAS:NLDAS_FORA0125_H.002:TMP2m&location=GEOM:POINT(
",
cities[i, "lon_nldas"],
",%20",
2013 Jan 19
2
Deformulation and R
Dear All,
I hope this is not too off-topic.
Essentially, I need to know if there is any R package which can help me
with a deformulation project.
Suppose e.g. that you know from a chemical analysis the fat, mineral,
vitamin, energy [and so on] content of a certain food product.
You also know the ingredients of this product (e.g. milk, lactose,
vegetable oil) and you know the chemical
2007 May 24
1
Basic connection between Mitel 3300 ICP and Asterisk (trixbox) - from a clueless newbie....
Hi all,
Our company has deployed a Mitel 3300 system (only about 2.5 years ago)
and we are experimenting with setting up Asterisk in our head office
(for business continuity, ie we have a bird flu epidemic and no-one can
come in, therefore use SIP softphones at home to co-ordinate activity)
and at a remote site in the Isle of Man (connected via 2Mbps SDSL)
Ideally we'd like anyone on either
2006 Feb 06
4
Relationship Question (STI)
My girlfriend and I have been dating for two years, and she just told me she
has an STI...
Actually, I currently have three different models, like: Dog, Whale, Monkey
They all have some similar attributes, but, they are unique enough to
break into their own models. I could use STI, but I think the table would
just have too many columns.
Now, I wanted to let the Dog''s, Whale''s
2010 Oct 28
2
R and Matlab
Dear Group,
I am looking for ways to use R and Matlab. Doing the data transformations in
R and using the data in Matlab to analyze with some pre-defined scripts.
Any good ways to transfer the data into matlab in its most recent version?
I tried using R.matlab but the writeMat output is not readable by Matlab.
I just need to output a data.frame and read it as is into matlab where I can
do any
2007 Mar 21
1
how to get "lsmeans"?
Dear all,
I search the mail list about this topic and learn that no simple way is available to get "lsmeans" in R as in SAS.
Dr.John Fox and Dr.Frank E Harrell have given very useful information about "lsmeans" topic.
Dr. Frank E Harrell suggests not to think about lsmeans, just to think about what predicted values wanted
and to use the predict
2007 Apr 05
1
p value for coefficients in multinomial model
Dear all,
1)how can I easily get p value for the coefficients of factors in a multinomial model?
2)why the p values for "type III" test with Anova are not identical to that from SAS?
for example:
A,B and C are categorical variables,but the proportions of each level in each categorical variables are not balance. Y is a nominal variables(>=3 categories);
To do
2004 Aug 05
1
ANOVA with repeated measurements (Dan Rajdl)
Dear R-users,
I have some human serum lipid concentrations (cholesterol, apoB ...),
each lipid was measured (in the same person) for 6 times in different
time points (start, 3 months, 6 months, 12 months, 18 months, 24
months). There were 2 groups of participants: one with a nutritional
intervention and the other without it. I would like to know, whether
lipid concentrations differ among different
2018 May 09
0
Seasonal weekly average
Hi Shakeel,
Assuming that you are starting with a bunch of dates:
# make a vector of character strings that can be converted to dates
rep_dates<-paste(sample(1:30,500,TRUE),sample(1:12,500,TRUE),
sample(2013:2017,500,TRUE),sep="/")
# if this isn't your format, change it
date_format<-"%d/%m/%Y"
# create a data frame with a column of dates
2023 Nov 03
1
I need to create new variables based on two numeric variables and one dichotomize conditional category variables.
Hello Everyone,
I have three variables: Waist circumference (WC), serum triglyceride (TG)
level and gender. Waist circumference and serum triglyceride is numeric and
gender (male and female) is categorical. From these three variables, I want
to calculate the "Lipid Accumulation Product (LAP) Index". The equation to
calculate LAP is different for male and females. I am giving both
2009 May 01
1
Last month on the Revolutions blog
I write about R every weekday at?http://blog.revolution-computing.com
. In case you missed them, here are some articles from the month of
April of particular interest to r-help subscribers. Thanks to everyone
who has been following the blog and sending me messages and/or leaving
comments -- it always brightens my day to hear from readers!
http://tinyurl.com/cy7x9a?(from April 1) announced the new