search for: 0.225

Displaying 20 results from an estimated 50 matches for "0.225".

Did you mean: 0.25
1998 Apr 27
1
R-beta: vectors in dataframe?
I have a file: x y z 0.025 0.025 1.65775 0.025 0.050 1.62602 0.025 0.075 1.63683 0.025 0.100 1.91847 0.025 0.125 2.00913 0.025 0.150 1.82222 0.025 0.175 1.70901 0.025 0.200 1.39759 0.025 0.225 1.39089 0.025 0.250 1.04762 If I read the file like this: data<-read.table("file.dat") How do I access the vectors x,y,z that are inside the dataframe data? I studied Venables and
2007 Mar 18
1
HELP...Running data
We are two french students and we have a problem concerning an exercize. We don't know how to resolve it. It would be fantastic if someone can help us. Thanks. Description: This study examined how the metabolic cost of locomotion varied with speed, stride frequency and body mass. Cost was determined by measuring oxygen consumption (?vo2?), analyzing the oxygen content in air inhaled and
2009 Jun 01
3
exporting data to csv file -problem with column names
I have a data.frame with three columns + the row number Sample MidPoint Count [1,] 1 0.025 1 [2,] 1 0.075 3 [3,] 1 0.125 15 [4,] 1 0.175 12 [5,] 1 0.225 5 [6,] 1 0.275 8 When I export the data to a csv the column names are shifted because the row numbers are being exported as well. Sample MidPoint
2010 Feb 04
2
help needed using t.test with factors
I am trying to use t.test on the following data: date type INTERVAL nCASES MTF SDF MTO SDO nFST MF nOBS MO MB BIASCV BIASEV ME MAE RMSE CRCF 2001-06-15 avn GE1.00 4385 0.246 0.300 1.502 0.556 1367 1.373 4385 1.502 1.471 0.285 0.164 -1.256 1.266 1.399 0.056 2001-06-15 avn
2011 Jul 17
3
?Accuracy of prop.test
I have just joined this list (and just started using R), so please excuse any etiquette breaches as I do not yet have a feel for how the list operates. I am in the process of teaching myself statistics using R as my utility as my ultimate goals cannot be satisfied by Excel or any of the plug-ins I could afford. I am currently looking at chap12 page 552 of Weiss's Introductory Statistics
2008 Aug 08
2
aggregate
Dear All- I have a dataset that is comprised of the following: doy yr mon day hr hgt1 hgt2 hgt3 co21 co22 co23 sig1 sig2 sig3 dif flag 244.02083 2005 09 01 00 2.6 9.5 17.8 375.665 373.737 373.227 3.698 1.107 0.963 -0.509 PRE 244.0625 2005 09 01 01 2.6 9.5 17.8 393.66 384.773 379.466 15.336 11.033 5.76 -5.307 PRE 244.10417 2005 09 01 02 2.6 9.5 17.8 411.162 397.866 387.755 6.835 5.61 6.728
2007 Oct 23
1
Strange dataframe behavior
Hello, I have a question regarding the following output: > database <- read.delim(file=path.input.file, header=TRUE, dec=".", sep="\t", na.strings = "#NV") > str(database) 'data.frame': 314 obs. of 13 variables: $ S : Factor w/ 314 levels "307073","400212",..: 147 72 299 137 162 62 189 236 134 307 ... $ A : Factor
2003 May 01
0
factanal
# I have a question about how factanal is calculating the regression factor # scores based on an oblique rotation (promax) of the factors. # # As is explained in the help file, regression factor scores are # obtained as # # hat f = Lambda' Sigma^-1 x # # However, according to Harman's "Modern Factor Analysis" (e.g. second # edition, pp. 351-352) the formula is # # hat f = Phi
2010 Dec 01
1
Poisson GLM warning message
Hi, I receive the following warning message when I run a poisson GLM in R: "glm.fit: fitted rates numerically 0 occurred" The model summary is shown below. The variable 'Species' consists of counts of different species ranging from 0 to 4. I suspect this may have something to do with the warning message but I'm not sure. Can anybody help? Thank you! Anna Call:
2003 Apr 24
1
write.table problem
Dear R helpers, I have been using the loadings function from the multiv library and I get the typical output (see below). When I try to export these results to a file using a write.table() I get the following error message "Error in as.data.frame.default(x[[i]], optional = TRUE) : can't coerce loadings into a data.frame" Any idea why write.table is doing that and any
2007 Dec 07
1
paradox about the degree of freedom in a logistic regression model
Dear all: "predict.glm" provides an example to perform logistic regression when the response variable is a tow-columned matrix. I find some paradox about the degree of freedom . > summary(budworm.lg) Call: glm(formula = SF ~ sex * ldose, family = binomial) Deviance Residuals: Min 1Q Median 3Q Max -1.39849 -0.32094 -0.07592 0.38220 1.10375
2012 Jul 31
1
kernlab kpca predict
Hi! The kernlab function kpca() mentions that new observations can be transformed by using predict. Theres also an example in the documentation, but as you can see i am getting an error there (As i do with my own data). I'm not sure whats wrong at the moment. I haven't any predict functions written by myself in the workspace either. I've tested it with using the matrix version and the
2002 Oct 09
3
proc mixed vs. lme
Dear All, Comparing linear mixed effect models in SAS and R, I found the following discrepancy: SAS R random statement random subj(program); random = ~ 1 | Subj -2*loglik 1420.8 1439.363 random effects variance(Intercept) 9.6033 9.604662
2010 Dec 21
2
please Help me on a repeated measures anova
I currently work on a draft of an aquatic bioassessment. The conditions tested are the following: ER river water T dechlorinated water control 0.5 + 0.5mg / L of malate T + 1 dechlorinated water control + 1g / L of malate T ED dechlorinated water control SED + ER + river water sediment SED ED + sediment + water dechlorinated. It is the result of AChE in muscle (fillet of fish). The production of
2008 Jul 20
5
[LLVMdev] qualitative comparison of correctness of llvm and gcc
Hi folks, We recently generated some data that seemed interesting enough to share here. This is a comparison between compilers that ignores the performance of the generated code and focuses only on compiler correctness. volatile checksum errors errors avr-gcc-3.4 1.879% 0.378% avr-gcc-4.1 0.037% 0.256% avr-gcc-4.2
2005 Mar 09
1
multiple comparisons for lme using multcomp
Dear R-help list, I would like to perform multiple comparisons for lme. Can you report to me if my way to is correct or not? Please, note that I am not nor a statistician nor a mathematician, so, some understandings are sometimes quite hard for me. According to the previous helps on the topic in R-help list May 2003 (please, see Torsten Hothorn advices) and books such as Venables &
2007 Jan 06
2
Bootstrapping Confidence Intervals for Medians
I apologize for this post. I am new to R (two days) and I have tried and tried to calculated confidence intervals for medians. Can someone help me? Here is my data: institution1 0.21 0.16 0.32 0.69 1.15 0.9 0.87 0.87 0.73 The first four observations compose group 1 and observations 5 through 9 compose group 2. I would like to create a bootstrapped 90% confidence interval on the difference of
2007 Apr 18
1
system call time increase when turning on CONFIG_PARAVIRT
Tim Chen wrote: > With CONFIG_PARAVIRT turned on, I've found that time invoking > system_call jumped up quite a lot. Using TCP streaming test as a > workload and running on 32-bit 2.6.20 kernel, system_call goes up from > 0.00025% all the way to 1.6% in the oprofile data. There is a drop of > about 4% in overall throughput for this particular workload. > > With
2008 Jan 18
0
forming a linear discriminant function from the output of lda()
Hello all- I am a relatively new user of R and am working through a graduate course in Statistics that uses Minitab, SAS and some Matlab. I like using R but am having some trouble lining up the output from lda() to that of the other programs' results. The dataset below is a modified set of wine data from the Pinot Noir data set as an illustration of the 2 group LDA scenario. Mo Ba
2009 Nov 09
0
Formula for calculating interaction terms in R
Hello - I am trying to figure out R's transformation for interaction terms in a linear regression. My simple background understanding is that interaction terms are generally calculated by multiplying the centred (0-mean) variables with each other and then doing the regression. However, in this regard I would have expected to see the same p-value when I calculate summary(lm(Y~A:B)) for A:B