Displaying 20 results from an estimated 2000 matches similar to: ""& not meaningful for factors""
2011 Jun 23
3
help- subtitles for multiple charts
Hello, I have a problem with my script. I don'y know how to apply subtitles.
I have 9 charts per page (for combination of 3 blocks and 3 treatments). I
want to have subtitles for this interaction (e.g. Block A Trt 1, Block A Trt
2, ...)
MBE$bt<- interaction(MBE$Block,MBE$trt)
par(mfrow=c(3,3))
for(i in unique(MBE$bt)){
ss <- MBE$bt==i
plot(MBE$Year[ss], MBE$DBH[ss])
2008 Oct 31
4
Help needed with Waterfall plot
Hi friends,
I need suggestions/directions on how to producing a waterfall plot for present extend of change in tumour size for a set of respondents in a study. Example of use of waterfall plot is in the following slides presented at ASCO 2007 by Axel Grothey. Link is
2008 May 28
5
"rbinom" not using probability of success right
I am trying to simulate a series of ones and zeros (1 or 0) and I am using "rbinom" but realizing that the number of successes expected is not accurate. Any advice out there.
This is the example:
N<-500
status<-rbinom(N, 1, prob = 0.15)
count<-sum(status)
15 percent of 500 should be 75 but what I obtain from the "count" variable is 77 that gives the probability of
2008 Jun 25
2
R alternative to SAS PROC REPORT
R has wonderful graphics but I am wondering whether there is anything in R to provide customization of tables like PROC Report does in SAS.
A Smile costs Nothing
But Rewards Everything
Happiness is not perfected until it is shared
-Jane Porter
[[alternative HTML version deleted]]
2008 May 19
2
Help on nested FOR loops
I am new to more radical programming in R. I am trying to write a nested 'for' loop to produce output that takes subscripts like:
for i taking values 1,2,3,4,5 and
j taking values 1,2,3
I want to output for a computation using the combination values of i and j a value x like this;
i j x
1 1 x11
1 2 x12
1 3 x13
2 1 x21
2 2 x22
2 3 x23
3 1 x31
3 2 x32
2009 Nov 22
1
Input file format to Anova from car package
Dear list member,
My question is related to input file format to an Anova from car package.
Here is an example of what I did:
My file format is like this (and I dislike the idea that I will need
to recode it):
Hormone day Block Treatment Plant Diameter High N.Leaves
SH 23 1 1 1 3.19 25.3 2
SH 23 1 1 2 3.42 5.5 1
SH 23 1 2 1 2.19 5.2 2
SH 23 1 2 2 2.17 7.6 2
CH 23 1 1 1 3.64 6.5 2
CH 23 1 1 2
2010 Sep 27
4
Fitting problem
Hi,
I have a function that generates a set of data but I am having problems
determining the parameters using the nls fitting procedure.
####
"MH"<-function(field,diameter,mu=10e-7,sig=0.1,Ms=100,chi=0){
#variables mu, sig, chi, Ms
#input: field and diameter
#all in CGS
rho <- 5
kb <- 1.38e-16
t <- 300
length.d<-length(diameter)
length.H<-length(field)
2008 Apr 15
2
glht with a glm using a Gamma distribution
Quick question about the usage of glht. I'm working with a data set
from an experiment where the response is bounded at 0 whose variance
increases with the mean, and is continuous. A Gamma error
distribution with a log link seemed like the logical choice, and so
I've modeled it as such.
However, when I use glht to look for differences between groups, I get
significant
2006 Aug 24
1
how to constrast with factorial experiment
Hello, R users,
I have two factors (treat, section) anova design experiment where
there are 3 replicates. The objective of the experiment is to test if
there is significant difference of yield between top (section 9 to 11)
and bottom (section 9 to 11) of the fruit tree under treatment. I
found that there are interaction between two factors. I wonder if I
can contrast means from levels of
2008 Oct 15
2
Network meta-analysis, varConstPower in nlme
Dear Thomas Lumley, and R-help list members,
I have read your article "Network meta-analysis for indirect treatment
comparisons" (Statist Med, 2002) with great interest. I found it very
helpful that you included the R code to replicate your analysis;
however, I have had a problem replicating your example and wondered if
you are able to give me a hint. When I use the code from the
2007 Mar 14
1
How to transform matrices to ANOVA input datasets?
Hello, R experts,
I have a list called dataHP which has 30 elements (m1, m2, ..., m30).
Each element is a 7x6 matrix holding yield data from two factors
experimental design, with treatment in column, position in row. For
instance, the element 20 is:
dataHP[[20]]
col1 col2 col3 trt1 trt2 trt3
[1,] 22.0 20.3 29.7 63.3 78.5 76.4
[2,]
2003 Jan 10
1
plot() and lines() multivariate problem
Dear list
I'm trying to solve the following problem since 2 days with no success.
Could someone help a newbie, please ?
I have a dependant variable which is diameter of shell, and two factors
which are level on the beach (3 levels 'B', 'E' and 'H') and a hydrodynamics
indice (3 levels 1, 2 and 3)
To study diameter variations according to these factors, I may use
2012 Jan 27
1
Confused with Student's sleep data description
I am confused whether Student's sleep data "show the effect of two
soporific drugs" or Control against Treatment (one drug). The reason
is the next:
> require(stats)
> data(sleep)
> attach(sleep)
> extra[group==1]
numeric(0)
> group
[1] Ctl Ctl Ctl Ctl Ctl Ctl Ctl Ctl Ctl Ctl Trt Trt Trt Trt Trt Trt Trt Trt Trt
[20] Trt
Levels: Ctl Trt
> sleep$group
[1] 1 1 1 1 1
2010 Dec 01
2
Lattice dotplots
Dear,
I have a dataset with 4 subjects (see ID in example), and 4 treatment (see
TRT in example) which are tested on 2 locations and in 3 blocs. By using
Lattice dotplot, I made a graph that shows the raw data per location and
per bloc. In that graph, I would like to have a reference line per bloc
that refers to the first treatment (T1). However, I can not find how to do
that.
I can make
2011 Apr 20
2
survexp with weights
Hello,
I probably have a syntax error in trying to generate an expected
survival curve from a weighted cox model, but I can't see it. I used
the help sample code to generate a weighted model, with the addition
of a "weights=albumin" argument (I only chose albumin because it had
no missing values, not because of any real relevance). Below are my
code with the resulting error
2008 Mar 05
1
Question on "assign(paste.."
Hello,
I'm having trouble in using "assign(paste ..." command . I could create
several dataframes following trinomial distribution using it but it could
not be used to check their row means of the created dataframe.
For example, the following works:
probTrt=matrix(0,4,3);
probTrt;
#malf, death, normal
probTrt[1,]=c(0.064,0.119,0.817);#for Trt 1
probTrt[2,]=c(0.053,0.125,0.823);#for
2005 Feb 22
1
Re: R-help Digest, Vol 24, Issue 22
You need to give the model formula that gave your output.
There are two sources of variation (at least), within and
between locations; though it looks as though your analysis
may have tried to account for this (but if so, the terms are
not laid out in a way that makes for ready interpretation.
The design is such (two locations) that you do not have
much of a check that effects are consistent over
2003 Jan 20
1
make check for R-1.6.2 on IBM AIX
Dear all,
The 'make check' step fails for the pacakge mva on IBM AIX.
The tail of the Rout log file looks like:
> for(factors in 2:4) print(update(Harman23.FA, factors = factors))
Call:
factanal(factors = factors, covmat = Harman23.cor)
Uniquenesses:
height arm.span forearm lower.leg weight
0.170 0.107 0.166
2012 May 29
2
use xyplot to plot mean and CI by groups
Dear R users,
I am trying to use xyplot to draw group mean and CI. The following is the
sample code. But I want:
1. Use different colors and symbols to draw individual points, CI and
the lines connect group means from different time points;
2. Add jitters to x axis to allow CIs not be overlapped
Could anyone modify the attached code to achieve this?
Thanks
library(lattice)
2011 Dec 11
2
multiple comparison of interaction of ANCOVA
Hi there,
The following data is obtained from a long-term experiments.
> mydata <- read.table(textConnection("
+ y year Trt
+ 9.37 1993 A
+ 8.21 1995 A
+ 8.11 1999 A
+ 7.22 2007 A
+ 7.81 2010 A
+ 10.85 1993 B
+ 12.83 1995 B
+ 13.21 1999 B
+ 13.70 2007 B
+ 15.15 2010 B
+ 5.69 1993 C
+ 5.76 1995 C
+ 6.39 1999