search for: 6.000000

Displaying 20 results from an estimated 45 matches for "6.000000".

Did you mean: 0.000000
2004 Apr 07
4
Problems with rlm
Dear all, When calling rlm with the following data, I get an error. (R v.1.8.1, WinXP Pro 2002 with service pack 1.) > d <- na.omit(data.frame(CPRATIO, HEIGHTZ, FAMILYID)) > c <- tapply(d$CPRATIO, d$FAMILYID, mean) > h <- tapply(d$HEIGHTZ, d$FAMILYID, mean) > c 1 2 3 6 7 9 10 11 6.000000 2.500000 3.250000
2011 Mar 11
5
How to calculate means for multiple variables in samples with different sizes
Hello R-helpers: I have data like this: sample replicate height weight age A 1.00 12.0 0.64 6.00 A 2.00 12.2 0.38 6.00 A 3.00 12.4 0.49 6.00 B 1.00 12.7 0.65 4.00 B 2.00 12.8 0.78 5.00 C 1.00 11.9 0.45 6.00 C 2.00 11.84 0.44 2.00 C 3.00 11.43 0.32 3.00 C 4.00 10.24 0.84 4.00 D
2010 Nov 01
1
floating-point issues with set_sort_by_relevance_then_value? (1.2.3, BM25 k1=0)
I am using BM25 with k1=0 and min_normlen=1 to get weights unaffected by document length and term frequency in the document (min_normlen=1 isn't necessary I guess) and am expecting single-term weights to be identical for all matches. I have added a document value to steer such general search queries and it works fine, except that for some search terms, I get results like:
2007 Jul 27
1
exporting character vector to text files
R-help, I have a character vector whose elements are the names of matrixes. Something like this: > test <- ls(pattern="Oki") [1] "aaOki" "aOki" "bOki" "c1Oki" "c2Oki" "c3Oki" "cOki" "dOki" "eOki" "fOki" "gprsOki" "hOki"
2007 Oct 17
1
How to save association rules generated by arules package
Hi, I have been able to generate association rules for Market Basket Analysis using the following codes: **************************************************************************** ******************************************* library("arules") rules <- read.csv("write1.csv",na.strings=c(".", "NA", "", "?"),header=TRUE)
2012 Feb 10
1
Need to aggregate large dataset by week...
Hi all, I have a large dataset with ~8600 observations that I want to compress to weekly means. There are 9 variables (columns), and I have already added a "week" column with 51 weeks. I have been looking at the functions: aggregate, tapply, apply, etc. and I am just not savvy enough with R to figure this out on my own, though I'm sure it's fairly easy. I also have the Dates
2004 Dec 29
1
Discrepancy between intervals.lme and coef.lme
I'm using R on Windows v2.0.1 with the nlme package (v3.1-53) and am finding some unexpected discrepancies in the output of intervals.lme and coef.lme. I've included a toy dataset at the end, but briefly, the data are longitudinal data from couples in marital therapy. Each spouse's relationship satisfaction is measured 4 times; I've fit both linear and quadratic models to the
2008 Oct 31
4
[ifelse] how to maintain a value from original matrix without probs?
Dear all, I have a matrix with positive and negative values. >From this I would like to produce 2 matrices: 1st - retaining positives and putting NA in other positions 2nd - retaining negatives and putting NA in other positions and then apply rowMeans for both. I am trying to use the function ifelse in the exemplified form: ifelse(A>0,A,NA) but by putting A as a 2nd parameter it
2008 Mar 03
1
Tapply for Group Specific Means and Proportions
UseRs, I am working on a dataset (see small example below) where individuals were followed on a specific date-time combo and multiple repeated measurements were taken (e.g., height in meters, behavior class in 2 letter code). Observation numbers varied between individual (ranging from 1 observation for each date-time combo to >50) I am trying to summarize the data into 1 row per
2010 Dec 08
1
I want to get smoothed splines by using the class gam
Hi all, I try to interpolate a data set in the form: time Erg 0.000000 48.650000 1.500000 56.080000 3.000000 38.330000 4.500000 49.650000 6.000000 61.390000 7.500000 51.250000 9.000000 50.450000 10.500000 55.110000 12.000000 61.120000 18.000000 61.260000 24.000000 62.670000 36.000000 63.670000 48.000000 74.880000 I want to get smoothed splines by using the class gam The first way I tried , was
2009 Aug 18
1
8.2 behaving weird on openSuSE 10.3 & 11.0
Listmates, I installed 8.2 on opensuse 10.3 from the opensuse XGL repository to replace 7.6. Compiz works with the nvidia 8600 GT card (512M of GDDR3) on the box, but the ctrl+alt <- -> switching behavior is really bad. In 7.6, the ctrl+alt <- -> desktop switching changed desktops crisply with just the slightest hint of the cube detected during the very short and nice switch.
2006 Aug 16
1
help about agnes
Hello. I have the following distance matrix between 8 points: [1,] 0.000000 3.162278 7.280110 8.544004 7.071068 9.899495 6.403124 8.062258 [2,] 3.162278 0.000000 5.000000 6.403124 4.472136 8.944272 6.082763 8.062258 [3,] 7.280110 5.000000 0.000000 1.414214 1.000000 5.000000 4.242641 5.830952 [4,] 8.544004 6.403124 1.414214 0.000000 2.236068 4.123106 4.472136 5.656854 [5,] 7.071068 4.472136
2009 Mar 20
1
Mean-replacing NAs in a 3d array
Hi all I have a 3d array containing missing values. > Xa , , 1 [,1] [,2] [1,] 1 3 [2,] NA 4 , , 2 [,1] [,2] [1,] 5 7 [2,] NA NA , , 3 [,1] [,2] [1,] 9 11 [2,] 10 12 I want to replace the missing values with the mean, but the mean of each 'page' in the array (wrong terminology I'm sure). So - for the array above - [2,1,2] and
2009 Jun 17
1
how to interpolate time series data with missingness
Hi all, I have a vector, most of which is missing. The data is always increasing, but may do so in jumps. I would like to interpolate the NAs with 'best guesses', using something like filter(), which doesn't work due to the NAs. Here is an example: > x <- c(2,3,NA,NA,NA,3.2,3.5,NA,NA,6,NA) > x [1] 2.0 3.0 NA NA NA 3.2 3.5 NA NA 6.0 NA I would like a function that
2024 Mar 29
2
Output of tapply function as data frame: Problem Fixed
Dear Rui, Thanks again for resolving this. I have already started using the version that works for me. But to clarify the second part, please let me paste the what I did and the error message: > set.seed(2024) > data <- data.frame( + Date = sample(seq(Sys.Date() - 5, Sys.Date(), by = "1 days"), 100L, + TRUE), + count = sample(10L, 100L, TRUE) + ) > > # coerce
2009 Aug 19
0
compiz Digest, Vol 42, Issue 3
I see no problem with your nvidia temps But, i have seen this exact behavior when 'indirect rendering' is enabled. On Wed, Aug 19, 2009 at 2:00 PM, <compiz-request at lists.freedesktop.org>wrote: > Send compiz mailing list submissions to > compiz at lists.freedesktop.org > > To subscribe or unsubscribe via the World Wide Web, visit >
2024 Mar 29
1
Output of tapply function as data frame: Problem Fixed
?s 01:43 de 29/03/2024, Ogbos Okike escreveu: > Dear Rui, > Thanks again for resolving this. I have already started using the version > that works for me. > > But to clarify the second part, please let me paste the what I did and the > error message: > >> set.seed(2024) >> data <- data.frame( > + Date = sample(seq(Sys.Date() - 5, Sys.Date(), by = "1
2024 Mar 27
1
Output of tapply function as data frame
Warm greetings to you all. Using the tapply function below: data<-read.table("FD1month",col.names = c("Dates","count")) x=data$count f<-factor(data$Dates) AB<- tapply(x,f,mean) I made a simple calculation. The result, stored in AB, is of the form below. But an effort to write AB to a file as a data frame fails. When I use the write table, it only produces
2006 Sep 08
3
newhidups with APC Smart-UPS 1500
Hello, I'm using the stable amd64 port of Debian Linux. I installed (the latest) nut-2.0.1-4 and nut-usb packages for utilizing an APC Smart-UPS 1500 USB. I tried "apcupsd" first but "nut" makes an even more sophisticated impression on me and has more security options. I had problems with the newhidups driver. It didn't find a device with matching VendorID. And
2009 May 24
1
Animal Morphology: Deriving Classification Equation with Linear Discriminat Analysis (lda)
Fellow R Users: I'm not extremely familiar with lda or R programming, but a recent editorial review of a manuscript submission has prompted a crash cousre. I am on this forum hoping I could solicit some much needed advice for deriving a classification equation. I have used three basic measurements in lda to predict two groups: male and female. I have a working model, low Wilk's lambda,