search for: 0.154

Displaying 20 results from an estimated 59 matches for "0.154".

Did you mean: 0.14
2008 May 28
5
"rbinom" not using probability of success right
I am trying to simulate a series of ones and zeros (1 or 0) and I am using "rbinom" but realizing that the number of successes expected is not accurate. Any advice out there. This is the example: N<-500 status<-rbinom(N, 1, prob = 0.15) count<-sum(status) 15 percent of 500 should be 75 but what I obtain from the "count" variable is 77 that gives the probability of
2007 Dec 11
3
matrix graph
Hi All, simple question: do you know how to graph the following object/matrix in a 'surface manner': [,1] [,2] [,3] [,4] [,5] [,6] [1,] -0.154 -0.065 0.129 0.637 0.780 0.221 [2,] 0.236 0.580 0.448 0.729 0.859 0.475 [3,] 0.401 0.506 0.310 0.650 0.822 0.448 [4,] 0.548 0.625 0.883 0.825 0.945 0.637 [5,] 0.544 0.746 0.823 0.877 0.861 0.642 [6,] 0.262 0.399
2010 Dec 21
2
please Help me on a repeated measures anova
I currently work on a draft of an aquatic bioassessment. The conditions tested are the following: ER river water T dechlorinated water control 0.5 + 0.5mg / L of malate T + 1 dechlorinated water control + 1g / L of malate T ED dechlorinated water control SED + ER + river water sediment SED ED + sediment + water dechlorinated. It is the result of AChE in muscle (fillet of fish). The production of
2009 Dec 21
3
Signif. codes
My question is about the "Signif. codes" and the p-value, specifically, the output when I run summary(nameofregression.lm) So you get this little key: Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1 And on a regression I ran, next to the intercept data, I get '***' Coefficients: > > Estimate Std. Error t value Pr(>|t|) > >
2009 Dec 08
1
Sort a data frame according to romans
R-help, I have a data frame: > mydata strata nh Nh Wh fh 1 I 10 26 0.048 0.385 2 II 32 84 0.154 0.381 3 III 16 42 0.077 0.381 4 IV 4 11 0.020 0.364 5 V 10 26 0.048 0.385 7 VII 64 168 0.309 0.381 8 VIII 49 129 0.237 0.380 9 IX 22 58 0.107 0.379 91 VI 0 0 0.000 0.000 and I wish to rearrange the data are sorted according to the roman
2010 Feb 17
2
extract the data that match
Hi r-users,   I would like to extract the data that match.  Attached is my data: I'm interested in matchind the value in column 'intg' with value in column 'rand_no' > cbind(z=z,intg=dd,rand_no = rr)             z  intg rand_no    [1,]  0.00 0.000   0.001    [2,]  0.01 0.000   0.002    [3,]  0.02 0.000   0.002    [4,]  0.03 0.000   0.003    [5,]  0.04 0.000   0.003    [6,] 
2012 Jun 13
2
asign variables in a "for" loop
Dear R-helpers, I'm stuck with a little problem that surely has an easy solution but I can't think of a way to solve it. I'd really appreciate any help you can offer me! I'll provide a small example. Given a dataframe data.txt that looks like this: ID freq Var Var_mean Ratio_mean Var_median Ratio_median Var_sum Ratio_min Var_max Ratio_max Var_min
2005 Nov 17
1
Principal Components Analysis (PR#8320)
Full_Name: Sahotra Sarkar Version: 2.2.0 OS: Windows XP Professional Submission from: (NULL) (146.6.130.180) The following two commands should give the same results for the eigenvectors but do not (there is a sign reversal for the first one): > summary(princomp(bumpus),loading = TRUE) Importance of components: Comp.1 Comp.2 Comp.3 Comp.4 Comp.5
2006 Jan 17
2
help with parsing multiple coxph() results
Dear All: I have a question on using coxph for multiple genes: I have written code to loop through all 22283 genes in the Hgu-133A and apply coxph on survival data. However, I don't know how to work with the result for each gene: survtest<-coxph(Surv(pcc.primary.stg.3.cox[,'fup_interval'],pcc.primary.stg.
2008 Jan 28
0
(no subject)
Hi all I am trying to generate a normal unbalanced data to estimate the coefficients of LM, LMM, GLM, and GLMM and their standard errors. Also, I am trying to estimate the variance components and their standard errors. Further, I am trying to use the likelihood ratio test to test H0: sigma^2_b = 0 (random effects variance component), and the t-test to test H0:mu=0 (intercept of the model Yij = mu
2007 Feb 19
2
Calculating the Sharpe ratio
Hi useRs, I am trying to calculate the Sharpe ratio with "sharpe" of the library "tseries". The documentation requires the univariate time series to be a portfolio's cumulated returns. In this case, the example given data(EuStockMarkets) dax <- log(EuStockMarkets[,"FTSE"]) is however not the cumulated returns but rather the daily returns of the FTSE
2009 Feb 23
1
why results from regression tree (rpart) are totally inconsistent with ordinary regression
Hi, In my analysis of impacts of insecticide-treated bednets on malaria, I look at the relationship between malaria incidence and mosquito behaviors. The condensed data set is copied here. Ordinary regression (lm) shows that Incidence was negatively related to Mortality. This makes sense because the latter reflected the strength of killing mosquitoes by insecticide-treated nets. Since the
2010 Dec 09
1
error in lrm( )
Dear Sir or Madam? I am a doctor of urology,and I am engaged in developing a nomogram of bladder cancer. May I ask for your help on below issue? I set up a dataset which include 317 cases. I got the Binary Logistic Regression model by SPSS.And then I try to reconstruct the model ?lrm(RECU~Complication+T.Num+T.Grade+Year+TS)? by R-Project,and try to internal validate the model through
2013 Jan 29
1
ccf (cross correlation function) problems
Hello everybody, I am sorry if my questions are too simple or not easily understandable. I’m not a native English speaker and this is my first analysis using this function. I have a problem with a cross correlation function and I would like to understand how I have to perform it in R. I have yearly data of an independent variable (x) from 1982 to 2010, and I also have yearly data of a variable
2018 May 15
2
Systemfit
OK, Let's try this again! Here is the reproducible script; it is long because I had to copy the panel dataset here. My question is related to systemfit; I don't know how to get the result for the entire panel. #Reproducible script Empdata<- read.csv("/Users/ngwinuiazenui/Documents/UPLOADemp.csv") View(Empdata) install.packages("systemfit")
2018 May 16
0
Systemfit
Sadly you failed to set your email program to send plain text and the data is corrupted at my end. I also think you need to reduce the size of the data set... the intent here is to increase your understanding, not debug your particular analysis. I will say that I am having a very challenging time understanding what you are trying to accomplish though. What are the equations that you think need
2018 May 15
0
Systemfit
... and the mailing list is picky about attachments... whatever you attached did not conform to the stringent requirements mentioned in the Posting Guide. Pasting the code right into the email is usually safest, though you DO have to post using plain text (as the Posting Guide indicates) or your code may get mangled by the automatic html format removal. On May 15, 2018 7:04:31 AM PDT, Bert Gunter
2019 Jan 03
1
--link-dest. Time to 'building file list' incrementing
I've been running rsync as a cygwin task on Windows Server 2008 for about two months now. I'm using the --link-dest option to do a daily 'snapshot' of the contents of a server containing about 10TB of data, about 13 million files, to a Linux based NAS server. Things started out great but I soon noticed that the time take to complete was slowly incrementing. It started at around
2012 Mar 02
2
Write performance in a replicated/distributed setup with KVM?
This has probably been discussed before, but since I'm new on the list I hope You have patience with me. I have a four brick distributed/replicated setup. The computers are multi-core 16GB memory and 2*2.0TB in raid1 SATA-disks locally. The nodes are connected by 1 GB ethernet. All nodes have glusterfs 3.3beta2 installed and they are running debian 6 64bit. The underlying filesystems are
2018 May 15
1
Systemfit
Unless there is good reason not to, always cc the list -- there are lots of smarter folks than I on it who can help. I may or may not have time to look at this. Hopefully someone else will. -- Bert Bert Gunter "The trouble with having an open mind is that people keep coming along and sticking things into it." -- Opus (aka Berkeley Breathed in his "Bloom County" comic strip