search for: 2.65

Displaying 20 results from an estimated 93 matches for "2.65".

Did you mean: 2.6
2012 Nov 14
1
indexing for Wilcoxon test
Hi everyone, I've been asked to run a number of Wilcoxon tests on some behavioral data (below is a sample). They want me to compare each trial to the first trial, considered the control trial. I know I can use brackets to index and select, for example, trial 1 vs. trial 4 (not exactly sure how to set that up either) but it would save me a ton of time if there was a way to do all the
2012 Nov 14
2
indexing for Wilcoxon test (take 2)
Hi everyone, I've been asked to run a number of Wilcoxon tests on some behavioral data (below is a sample). They want me to compare each trial to the first trial, considered the control trial. I know I can use brackets to index and select, for example, trial 1 vs. trial 4 (not exactly sure how to set that up either) but it would save me a ton of time if there was a way to do all the
2004 Feb 19
1
reshape direction=wide
Hello I am reshaping a data.frame bids --> reshaped as shown below. I thought this should be possible with a single invocation of reshape, but the only way I came up with is reshaping subsets for each keyword and then joining them together. Does anyone have an idea how to solve this in a more elegant way? Efficiency is a concern as the datasets are very large. Is there a way to specify
2009 Apr 04
2
data.frame, converting row data to columns
I have a data frame something like: name wrist nLevel emot 1 4094 3.34 1 frustrated 2 4094 3.94 1 frustrated 3 4094 NA 1 frustrated 4 4094 3.51
2009 Dec 28
2
Modified R Code
Dear R helpers,   I have following input files. (Actually they are more than 10 rates but here i am considering only 2 rates to write my problem)   rate1.csv min1        max1            min2          max2          min3           max3 1.05        1.30               1.30          1.65             1.65          1.99   rate2.csv min1        max1            min2          max2          min3          
2012 Mar 03
3
How to read this data properly?
Dear all, I have been given a data something like below: Dat = "2 3 28.3 3.05 8 3 3 22.5 1.55 0 1 1 26.0 2.30 9 3 3 24.8 2.10 0 3 3 26.0 2.60 4 2 3 23.8 2.10 0 3 2 24.7 1.90 0 2 1 23.7 1.95 0 3 3 25.6 2.15 0 3 3 24.3 2.15 0 2 3 25.8 2.65 0 2 3 28.2 3.05 11 4 2 21.0 1.85 0 2 1 26.0 2.30 14 1 1 27.1 2.95 8 2 3 25.2 2.00 1 2 3 29.0 3.00 1 4 3 24.7 2.20 0 2 3 27.4 2.70 5 2 2 23.2 1.95
2013 Feb 26
2
Help on a code
Dear List,   I have the following code:   x <-c(0, 13.8333333333333, 38.1666666666667, 62.1666666666667, 85.9166666666667, 108.916666666667) y <-c(1.77, 2.39, 3, 2.65, 2.62, 1.8) Interpolated <- approx(x, y,xout=0:tail(x, n=1),method="linear") plot(Interpolated)   in this code x is time in hours (cumulative), and y is a biological variable. I am using linear interpolation
2010 Feb 19
5
autoconf error with Wine 1.1.39
I have got this error when trying to build Wine 1.1.39 in Fedora 12 Code: + autoreconf configure:4993: error: possibly undefined macro: AS_VAR_APPEND If this token and others are legitimate, please use m4_pattern_allow. See the Autoconf documentation. autoreconf: /usr/bin/autoconf failed with exit status: 1 In Fedora 12 installed autoconf-2.63.
2012 Jul 30
2
distance matrix and hclustering
Dear R Users,i am very new to R. I want your help on an issue regarding distance matrix and cluster analysis i had discharge data of 4 rivers(a,b,c,d) in 4 vectors each having 364 values > dput(qmu)structure(list(a = c(0.26, 0.25, 0.25, 0.25, 0.24, 0.23, 0.22, 0.21, 0.21, 0.21, 0.2, 0.19, 0.19, 0.19, 0.19, 0.18, 0.18, 0.18, 0.17, 0.17, 0.17, 0.17, 0.17, 0.17, 0.17, 0.17, 0.17, 0.17, 0.17, 0.17,
2004 May 21
2
bars with sd
I need some help for a curious question of a friend of mine. She usually does some experiments (3-5 repeats for each exp) and then she calculates mean and standard deviation. In microsoft excel she writes something like the following sample mean sd a 1.25 0.35 b 2.65 0.65 c 3.45 0.50 She can do a vertical barplot graph just giving mean value and specifying the
2002 Oct 17
4
Posix Problem, difftime
I am having a series of problems using date time data that has been converted into a POSIXt and POSIXlt classes. I have hourly time series data from 1900 that has been converted from text data. I assume most of my problems come from a mis-underdanding of the POSIX class. My matrix named (aa) for this year is approx 8700 by 4. When I try to calculate the length of posit column ( which is the
2012 Jun 20
2
[LLVMdev] Exception handling slowdown?
Did something change with exception handling recently? A bunch of lit bots are showing slower compile times for many tests. Ciao, Duncan. On 20/06/12 07:53, llvm-testresults at cs.uiuc.edu wrote: > > lab-mini-03__O0-g__clang_DEV__x86_64 test results > <http://llvm.org/perf/db_default/v4/nts/1283?compare_to=1278&baseline=999> > > Run Order Start Time Duration >
2009 Oct 23
2
interpretation of RCS 'coefs' and 'knots'
Hi, I have fit a series of ols() models, by group, in this manner: l <- ols(y ~ rcs(x, 4)) ... where the series of 'x' values in each group is the same, however knots are not always identical between groups. The result is a table of 'coefs' derived from the ols objects, by group: group Intercept top top' top'' 1 6.864 0.01 2.241 -2.65
2012 Jun 25
0
[LLVMdev] Exception handling slowdown?
Nothing that I'm aware of has changed with EH. Is it possible to bisect the problem? -bw On Jun 20, 2012, at 12:38 AM, Duncan Sands <baldrick at free.fr> wrote: > Did something change with exception handling recently? A bunch of lit bots are > showing slower compile times for many tests. > > Ciao, Duncan. > > On 20/06/12 07:53, llvm-testresults at cs.uiuc.edu
2013 Jul 22
3
2.2.4 + metadata plugin: autoconf failed
Hello, I can compile metadata plugin using debian squeeze + wheezy. But build on suse enterprise server 9,10 and 11 failed. The metadata plugin require autoconf-2.65 which i too new. On the other side I can build the dovecot-2.2.4 and pigeonhole-0.4.0 plugin without problems: dovecot require autoconf-2.59 and pigeonhole does not require any specific autoconf version. I asked the authors of the
2012 Jul 05
2
[LLVMdev] Exception handling slowdown?
Hi Bill, > Nothing that I'm aware of has changed with EH. Is it possible to bisect the problem? I don't see any relevant LLVM changes, so I guess clang C++ compilation slowed down due to some clang changes. I'm not going to investigate this. Ciao, Duncan. > > -bw > > On Jun 20, 2012, at 12:38 AM, Duncan Sands <baldrick at free.fr> wrote: > >> Did
2006 Feb 21
1
call row names
Hi R users. I have a table like that: table var A1 A2 A3 v1 41203 3.69 2.31 v2 20577 4.51 8.60 v3 20625 2.87 3.50 v4 6115 8.92 2.97 v5 3160 1.49 2.21 v6 2954 2.62 5.98 v7 4731 1.83 7.53 v8 2435 7.68 3.50 v9 2296 3.03 4.84 v10 6153 1.06 4.28 v11 3157 1.07 1.15 v12 2996 1.06 1.01 v13 6084 2.65 2.63 v14 3115 2.42
2007 Jun 05
1
logit model interpretation
Hello everyone I appologize for my lack of experience in statistical methods. I am an R user begginer and I am running a logit model using "zelig" and "pcse" packages. I will go to the point and is that Im having problems with interpreting the results of my models.. It is really simple (I guess for the most advanced scholars) however I really dont understand how to interpret
2007 Sep 24
2
calculating/plotting error ellipses
hello, sorry for posting what may be a simple question: i do have a matrix of coordinates (positional judgments, see below) and now want to calculate and plot the corresponding error ellipse. can anyone help me with the exact steps/syntax? > xyDat X Y 1 -0.49 -2.13 2 0.91 0.48 3 0.20 -2.80 4 -0.76 -3.23 5 -0.36 2.50 6 1.38 1.24 7
2011 Jan 31
0
Forest Plot Text
Hello All, I am having a problem creating text for my forest plot using the "forestplot" function in the rmeta package. I think my problem is that I have too many columns of text, so I was wondering if there is a way to shrink my text or to change the default setting. Here is an example: #### forest_mean<- as.numeric(c(NA, 1.00, 2.00, NA, 0.75, 1.10)) forest_lower<-