similar to: Again question about filter()

Displaying 20 results from an estimated 80 matches similar to: "Again question about filter()"

2006 Apr 18
2
what happen?
Hi! I want to choose part of the data. But it has a question. > attach(PIKO) > PIKO[1:5,] code species dbh x y tag status branch 92 10142 PIKO 38.9 6.05 12.81 10165 alive 0 109 10213 PIKO 41.0 6.71 26.21 10202 alive 0 135 10222 PIKO 48.3 18.98 21.28 10214 alive 0 146 10223 PIKO 47.0 19.45 22.83 10224 alive 0 147 10223 PIKO 20.0 19.74
2008 Mar 14
1
Lme does not work without a random effect (UNCLASSIFIED)
Classification: UNCLASSIFIED Caveats: NONE Dear R users, I'm interested in finding a random effect of the Block in the data shown below, but 'lme' does not work without the random effect. I'm not sure how to group the data without continuous value which is shown in the error message at the bottom line. If I use 'aov' with Error(Block), is there a test method comparing
2013 Jan 13
1
Extract data in word pad
Dear r-users,   I have saved data in word pad.  I would like to extract certain part of data only, for example   19710629 080000(PARTIAL) 39.3 at interval beginning 19701126 010326 19720629 080000(PARTIAL) 33.8 at interval beginning 19720517 144507 19730629 080000(PARTIAL) 32.2 at interval beginning 19720910 135747   and so on...   The original data set are given below:      
2013 Jan 11
2
Extract data
Dear R users, I just join this forum a few minutes ago. My friend recommend this forum to me.   I have data in excel csv. My problem is I would like to extract some number, for example as highlighted below;   Year Ending           Maximum Value (Rain mm)  Gap from 19710509 090000 to 19710607 080000 of   28.96 Days 19710629 080000(PARTIAL)       39.3 at interval beginning 19701126 010326  Gap from
2006 Nov 18
1
deriv when one term is indexed
Hi, I'm fitting a standard nonlinear model to the luminances measured from the red, green and blue guns of a TV display, using nls. The call is: dd.nls <- nls(Lum ~ Blev + beta[Gun] * GL^gamm, data = dd, start = st) where st was initally estimated using optim() st $Blev [1] -0.06551802 $beta [1] 1.509686e-05 4.555250e-05 7.322720e-06 $gamm [1] 2.511870 This works fine but I
2010 Jan 04
1
log-normal overlay
Hello, Using the following lines of code, I created the following graph:
2010 Jan 04
1
log normal overlay
Hello, Using the following lines of code, I created the following graph:
2017 Sep 22
1
Treating NA in timeSeries package
Dear All, I am facing problem with NA treatment in my financial time series data. # data reading aluminum = read.csv(file="alu.csv", header=T, sep=",") fut = aluminum [,2] spt = aluminum [,3] # Missing Value Treatment (Linear Interpolation) spt = interpNA(spt, method = c("linear")) fut = interpNA(fut, method = c("linear")) fut=fut[,1] spt =spt[,1]
2011 May 16
2
wireframe advice - with reproducible code
Dear List, i am trying to produce a 3d plot using wireframe using the code: wireframe(Residuals_FD ~ Elevation * Temperature, data = data2, scales = list(arrows = FALSE), drape = TRUE, colorkey = TRUE) As you can see when the code (using the data below) is run the plot area is set-up correctly but the actual surface is missing? Any help would be greatly appreciated. Chris #data Elevation
2010 May 26
25
Extremely slow start up with ruby 1.9.1 vs 1.8.7
I created a fresh new rails app, no gems, no nothing. Completely unmodified. Doing anything that initializes rails takes 2 - 3 times longer in ruby 1.9.1 than it does in 1.8.7. Obviously with a fresh rails app, this is only 2 - 3 seconds extra. But with a decent sized app, this can take up to 25 seconds for rails to initialize, vs the 12 seconds in 1.8.7. I''ve been racking my brain on
2012 Apr 17
1
Test-Predict R survival analysis
Hi, I'm trying to use the R Survival analysis on a windows 7 system. The input data format is described at the end of this mail. 1/ I tried to perform a survival analysis including stratified variables using the following formula. cox.xtab_miR=coxph(Surv(time, status) ~ miR + strata(sex,nbligne, age), data=matrix) and obtain the following error message Warning message: In fitter(X, Y,
2012 May 04
0
[LLVMdev] TestSuite reports automatic comparison
Hi guys! Is there any automatic way to generate the *comparison* report between two test-reports? I mean the tool that automatically generates something like this: 1 = some-file-at-left.json 2 = some-file-at-right.json 1:compile | 1: exec | 2: compile | 2: exec | %compile | %exec Burg | 1.21 | 0 | 1.13 | 0 | -6.7% | - ldecod | 6.71 | 0.06 | 6.88
2006 Dec 14
1
subset question
I have a data set p1982, its structure is the following Then I take 20 observations from this dataset, and assign to pr. in p1982, p has 1982 levels, in dataset pr, p should have 1 levels. But I do str(pr), it shows that p still has 1982 levels. also for these > pr$aa [1] ARG THR ASP CYS TYR ASN VAL ASN ARG ILE ASP THR THR ALA SER CYS LYS THR ALA LYS Levels: ALA ARG ASN ASP CYS GLN
2008 Mar 17
0
Summary Regard with Lme does not work without a random effect (UN CLASSIFIED)
Classification: UNCLASSIFIED Caveats: NONE R users, This is a summary for the responses I got from Sandra Dorai-Raj and Simon Blomberg followed by my question. Sandra suggested using lm instead of lme for a model without a random effect, and Simon suggested anova.lme instead of anova.lm for comparing two models with and without a random effect. For the anova test, both anova.lme and anova
2008 Sep 08
0
Poisson Distribution - Chi Square Test for Goodness of Fit
Dear R-help,   Chi Square Test for Goodness of Fit     Problem Faced :   I have got a discrete data as given below (R script)   No_of_Frauds <-c 1,1,1,1,1,1,1,1,1,2,1,1,1,1,1,1,2,1,2,2,2,1,1,2,1,1,1,1,4,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,5,1,2,1,1,1,1,1,1,1,3,2,1,1,1,2,1,1,2,1,1,1,1,1,2,1,3,1,2,1,2,14,2,1,1,38,3,3,2,44,1,4,1,4,1,2,2,1,3)   I am trying to fit
2008 Aug 29
0
Problem with Poisson - Chi Square Goodness of Fit Test - New Mail
Dear R-help, ? ? Chi Square Test for Goodness of Fit ? I have got a discrete data as given below (R script) ? No_of_Frauds<-c(1,1,1,1,1,1,1,1,1,2,1,1,1,1,1,1,2,1,2,2,2,1,1,2,1,1,1,1,4,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,5,1,2,1,1,1,1,1,1,1,3,2,1,1,1,2,1,1,2,1,1,1,1,1,2,1,3,1,2,1,2,14,2,1,1,38,3,3,2,44,1,4,1,4,1,2,2,1,3) ? I am trying to fit Poisson distribution to
2008 Aug 29
0
Poisson Distribution - problem with Chi Square Goodness of Fit test
Chi Square Test for Goodness of Fit   I have got a discrete data as given below (R script)   No_of_Frauds<-c(1,1,1,1,1,1,1,1,1,2,1,1,1,1,1,1,2,1,2,2,2,1,1,2,1,1,1,1,4,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,5,1,2,1,1,1,1,1,1,1,3,2,1,1,1,2,1,1,2,1,1,1,1,1,2,1,3,1,2,1,2,14,2,1,1,38,3,3,2,44,1,4,1,4,1,2,2,1,3)   I am trying to fit Poisson distribution to this data using R.  
2008 Aug 29
0
Please ignore earlier mail - [ Poisson - Chi Square test for Goodness of Fit]
Dear R-help, ? ? Chi Square Test for Goodness of Fit ? I have got a discrete data as given below (R script) ? No_of_Frauds<-c(1,1,1,1,1,1,1,1,1,2,1,1,1,1,1,1,2,1,2,2,2,1,1,2,1,1,1,1,4,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,5,1,2,1,1,1,1,1,1,1,3,2,1,1,1,2,1,1,2,1,1,1,1,1,2,1,3,1,2,1,2,14,2,1,1,38,3,3,2,44,1,4,1,4,1,2,2,1,3) ? I am trying to fit Poisson
2008 Oct 05
1
io writes very slow when using vmware server
We are struggling with a strange problem. When we have some VMWare clients running (mostly MS windows clients), than the IO-write performance on the host becomes very bad. The guest os's do not do anything, just having them started, sitting at the login prompt, is enough to trigger the problem. The host has plenty of 4G of RAM, and all clients fit easily into the space. The disksystem is a
2009 Oct 20
0
[LLVMdev] 2.6 pre-release2 ready for testing
Hi Tanya, > 1) Compile llvm from source and untar the llvm-test in the projects > directory (name it llvm-test or test-suite). Choose to use a > pre-compiled llvm-gcc or re-compile it yourself. I compiled llvm and llvm-gcc with separate objects directories. Platform is x86_64-linux-gnu. > 2) Run make check, report any failures (FAIL or unexpected pass). Note > that you need to