suiming47
2010-Nov-24 02:05 UTC
[R] 4. Rexcel (Luis Felipe Parra)-how to run a code from excel
Hi Louis, It's simple to run a r script from the excel spreadsheet. Just write your code, source("C:\\Quantil Aplicativos\\Genercauca\\BackwardSelectionNC.r"), into a cell of a workingsheet. Then right-click the cell and select "run code" in the pop-up menu. Hope this will help you. Best, Bernard -----????----- ???: r-help-bounces at r-project.org [mailto:r-help-bounces at r-project.org] ?? r-help-request at r-project.org ????: 2010?11?23? 19:00 ???: r-help at r-project.org ??: R-help Digest, Vol 93, Issue 23 Send R-help mailing list submissions to r-help at r-project.org To subscribe or unsubscribe via the World Wide Web, visit https://stat.ethz.ch/mailman/listinfo/r-help or, via email, send a message with subject or body 'help' to r-help-request at r-project.org You can reach the person managing the list at r-help-owner at r-project.org When replying, please edit your Subject line so it is more specific than "Re: Contents of R-help digest..." Today's Topics: 1. Re: "unexpected numeric constant" while reading tab delimited csv file (Mike Marchywka) 2. how to sample lowess/loess into matrix ? (madr) 3. Re: "unexpected numeric constant" while reading tab delimited csv file (Duncan Murdoch) 4. Rexcel (Luis Felipe Parra) 5. how to change number of characters per line for print() to sink()? (Nevil Amos) 6. Invitation ? se connecter sur LinkedIn (Aline Uwimana via LinkedIn) 7. Re: How to combine Date and time in one column (Gabor Grothendieck) 8. Tinn-R 2.3.7.0 released (Jose Claudio Faria) 9. Rexcel (Luis Felipe Parra) 10. Re: how to change number of characters per line for print() to sink()? (jim holtman) 11. Re: boxplot: reverse y-axis order (S Ellison) 12. plot inside function does not work (Alaios) 13. Re: how to sample lowess/loess into matrix ? (madr) 14. Re: plot inside function does not work (Claudia Beleites) 15. Re: Rexcel (Stephen Liu) 16. Re: plot inside function does not work (Alaios) 17. Invitation ? se connecter sur LinkedIn (Yann Lancien via LinkedIn) 18. Prob with merge (Joel) 19. Re: Making a line in a legend shorter (Peter Ehlers) 20. Re: Prob with merge (ONKELINX, Thierry) 21. Re: calculating martingale residual on new data using "predict.coxph" (Terry Therneau) 22. Re: Prob with merge (Joel) 23. What if geoRglm results showed that a non-spacial model fits? (Jimmy Martina) 24. Help with plotting kohonen maps (Stella Pachidi) 25. Re: ?summaryRprof running at 100% cpu for one hour ... (Kjetil Halvorsen) 26. Re: ?summaryRprof running at 100% cpu for one hour ... (Mike Marchywka) 27. hierarchical mixture modelling (Mike Lawrence) 28. Check for is.object (Santosh Srinivas) 29. Fast Two-Dimensional Optimization (Wonsang You) 30. Re: plotting a timeline (?ukasz R?c?awowicz) 31. Re: memory profiling (Patrick Leyshock) 32. Re: ?summaryRprof running at 100% cpu for one hour ... (Kjetil Halvorsen) 33. Re: An empty grey diagram (Martin Maechler) 34. Re: ?summaryRprof running at 100% cpu for one hour ... (Mike Marchywka) 35. Re: Check for is.object (Jonathan P Daily) 36. Re: solve nonlinear equation using BBsolve (Berend Hasselman) 37. Re: Alternatives to image(...) and filled.contour(...) for 2-D filled Plots (Ista Zahn) 38. Re: statistical test for comparison of two classifications (nominal) (Matt Shotwell) 39. Re: Check for is.object (Phil Spector) 40. Re: "negative alpha" or custom gradient colors of data dots in scatterplot ? (Ista Zahn) 41. Problem setting the number of digits in xtable (wphantomfr) 42. RCurl : All connection are used ? (omerle) 43. sm.ancova graphic (Lucia Ca?as) 44. Plotting a cloud/fog of variable density in rgl (JiHO) 45. Re: RCurl : All connection are used ? (Mike Marchywka) 46. Re: Ordeing Zoo object (Manta) 47. Re: Plotting a cloud/fog of variable density in rgl (Duncan Murdoch) 48. Re: How to produce a graph of glms in R? (Greg Snow) 49. Re: Plotting a cloud/fog of variable density in rgl (Mike Marchywka) 50. Re: data acquisition with R? (B.-Markus Schuller) 51. Is it possible to make a matrix to start at row 0? (bogdanno) 52. Re: Rexcel (csrabak) 53. Re: question about constraint minimization (dhacademic at gmail.com) 54. Re: Lost in POSIX (Dimitri Shvorob) 55. Re: Find in R and R books (Georg Otto) 56. Re: using rpart with a tree misclassification condition (meytar) 57. Re: memory profiling (Patrick Leyshock) 58. Some questione about plot (romzero) 59. how do remove those predictor which have p value greater than 0.05 in GLM? (shubha) 60. Re: how to apply sample function to each row of a data frame? (wangwallace) 61. Re: arima (tomreilly) 62. Wait for user input with readline() (Nathan Miller) 63. Re: txtProgressBar strange behavior in R 2.12.0 (Viechtbauer Wolfgang (STAT)) 64. How to call web service in R (??) 65. save a regression model that can be used later (Ni, Melody Zhifang) 66. Re: question about constraint minimization (Ravi Varadhan) 67. Re: Is it possible to make a matrix to start at row 0? (Joshua Wiley) 68. Re: Is it possible to make a matrix to start at row 0? (Bert Gunter) 69. Re: RGoogleDocs stopped working (Harlan Harris) 70. Re: sm.ancova graphic (Peter Ehlers) 71. Re: Wait for user input with readline() (Joshua Wiley) 72. Re: Ordeing Zoo object (Gabor Grothendieck) 73. Re: Is it possible to make a matrix to start at row 0? (Ben Bolker) 74. Re: save a regression model that can be used later (David Winsemius) 75. how to round only one column of a matrix ? (madr) 76. Re: Is it possible to make a matrix to start at row 0? (baptiste auguie) 77. Re: how to round only one column of a matrix ? (Phil Spector) 78. Re: How to call web service in R (Steve Lianoglou) 79. aggregate a Date column does not work? (Tan, Richard) 80. R2WinBUGS help (bmiddle) 81. Re: FW: help with time Series regression please (tomreilly) 82. Re: Is it possible to make a matrix to start at row 0? (Mike Marchywka) 83. Re: aggregate a Date column does not work? (David Winsemius) 84. I need a very specific unique like function and I don't know even how to properly call this (madr) 85. Re: aggregate a Date column does not work? (Tan, Richard) 86. R package "kernlab" can not be properly loaded (Xiaoqi Cui) 87. Re: Rexcel (Erich Neuwirth) 88. Re: aggregate a Date column does not work? (David Winsemius) 89. Re: aggregate a Date column does not work? (Tan, Richard) 90. Re: aggregate a Date column does not work? (Gabor Grothendieck) 91. Re: "negative alpha" or custom gradient colors of data dots in... (madr) 92. Re: how do remove those predictor which have p value greater than 0.05 in GLM? (Frank Harrell) 93. Help: Standard errors arima (lucia) 94. Re: Help: Standard errors arima (David Winsemius) 95. Re: R2WinBUGS help (Uwe Ligges) 96. Re: R package "kernlab" can not be properly loaded (David Winsemius) 97. Re: R package "kernlab" can not be properly loaded (Uwe Ligges) 98. Re: I need a very specific unique like function and I don't know even how to properly call this (Ista Zahn) 99. Re: R2WinBUGS help (bmiddle) 100. Re: how do remove those predictor which have p value greater than 0.05 in GLM? (shubha) 101. Re: Help: Standard errors arima (Dennis Murphy) 102. Probit Analysis: Confidence Interval for the LD50 using Fieller's and Heterogeneity (UNCLASSIFIED) (Kenney, Colleen T CTR USA AMC) 103. how to calculate derivative (Yogesh Tiwari) 104. Re: cpgram: access data, confidence bands (David Scott) 105. Re: Probit Analysis: Confidence Interval for the LD50 using Fieller's and Heterogeneity (UNCLASSIFIED) (David Winsemius) 106. Re: Probit Analysis: Confidence Interval for the LD50 using Fieller's and Heterogeneity (UNCLASSIFIED) (Dennis Murphy) 107. empity value in colnames (M.Ribeiro) 108. Re: ?summaryRprof running at 100% cpu for one hour ... (Kjetil Halvorsen) 109. Re: empity value in colnames (David Winsemius) 110. Re: ?summaryRprof running at 100% cpu for one hour ... (Mike Marchywka) 111. Sporadic errors when training models using CARET (Kendric Wang) 112. Re: I need a very specific unique like function and I don't know even how to properly call this (Phil Spector) 113. Re: how do remove those predictor which have p value greater than 0.05 in GLM? (David Winsemius) 114. Re: plotting a timeline (Peter Ehlers) 115. Gap between graph and axis (Sebastian Rudnick) 116. Re: Gap between graph and axis (Bill.Venables at csiro.au) 117. R on Androids? (Erin Hodgess) 118. Re: how to loop through variables in R? (watashi at post.com) 119. How to start default browser on R (Stephen Liu) 120. question on "uniCox" (Shi, Tao) 121. Re: Find in R and R books (Spencer Graves) 122. Re: calculating martingale residual on new data using "predict.coxph" (Shi, Tao) 123. Re: calculating martingale residual on new data using (Shi, Tao) 124. Re: calculating martingale residual on new data using "predict.coxph" (Shi, Tao) 125. Re: Find in R and R books (Mike Marchywka) 126. Re: Fast Two-Dimensional Optimization (Ben Bolker) 127. Re: how to loop through variables in R? (watashi at post.com) 128. Re: How to start default browser on R (David Scott) 129. Re: Find in R and R books (Spencer Graves) 130. using the "apply" method for functions with multiple inputs (joeponzio) 131. permalgorithm (Kere Klein) 132. Re: using the "apply" method for functions with multiple inputs (David Winsemius) 133. Re: How to start default browser on R (Stephen Liu) 134. Re: How to start default browser on R (Ista Zahn) 135. Re: how to calculate derivative (Ravi Varadhan) 136. Re: How to start default browser on R (Stephen Liu) 137. Re: How to start default browser on R (David Scott) 138. Re: how to calculate derivative (Spencer Graves) 139. Re: Is it possible to make a matrix to start at row 0? (David Stoffer) 140. overlay histograms on map at map coordinates (Steve Bellan) 141. Re: Kalman filter (David Stoffer) 142. (no subject) (Mari Pesek) 143. factorial ANOVA for block/split-plot design (Mari Pesek) 144. Re: factorial ANOVA for block/split-plot design (RICHARD M. HEIBERGER) 145. Re: question about constraint minimization (dhacademic at gmail.com) 146. Explained GLZ model variation (Shai) 147. Lattice and Quartz (Sarah Berry) 148. compare GLM coefficients (Kayce anderson) 149. Re: compare GLM coefficients (Michael Bedward) 150. Calculating correlation (gireesh bogu) 151. Re: Some questione about plot (Jim Lemon) 152. Re: How to start default browser on R (Stephen Liu) 153. Re: Calculating correlation (Tal Galili) 154. About available datasets on PC (Stephen Liu) 155. More detail in chart axis? (Noah Silverman) 156. Error: cannot allocate vector of size x Gb (64-bit ... yet again) (derek eder) 157. Re: About available datasets on PC (Jeff Newmiller) 158. Re: More detail in chart axis? (Jim Lemon) 159. Re: how to get rid of unused space on all 4 borders in plot() render (Petr PIKAL) 160. Re: how to loop through variables in R? (Ivan Calandra) 161. Odp: save a regression model that can be used later (Petr PIKAL) 162. Re: Lost in POSIX (Jeff Newmiller) 163. Re: About available datasets on PC (Stephen Liu) 164. Re: Calculating correlation (Tal Galili) ---------------------------------------------------------------------- Message: 1 Date: Mon, 22 Nov 2010 06:29:34 -0500 From: Mike Marchywka <marchywka at hotmail.com> To: <jdnewmil at dcn.davis.ca.us>, <madrazel at interia.pl> Cc: r-help at r-project.org Subject: Re: [R] "unexpected numeric constant" while reading tab delimited csv file Message-ID: <BLU113-W14BF3EA0E3821E3C0CA8AEBE3D0 at phx.gbl> Content-Type: text/plain; charset="iso-8859-1" ----------------------------------------> Date: Mon, 22 Nov 2010 01:57:54 -0800 > From: jdnewmil at dcn.davis.ca.us > To: madrazel at interia.pl > CC: r-help at r-project.org > Subject: Re: [R] "unexpected numeric constant" while reading tab delimitedcsv file> > madr wrote: > > my csv file is very simple - just one line for purpose of this test: > > 0{TAB}0 > > > > and read function is this: > > csvdata = read.csv(file="d:/s/test.csv",head=FALSE,sep="\t") > > > > then error comes: > > > > Error in source("d:/test.csv") : > > d:/test.csv:1:9: unexpected numeric constant > > 1: 0 0 > > > > > > but when I change delimiter to ; (colon) then error not shows up anymore > > > You seem to be referencing two different files somehow... one in the > root directory of your drive D, and the other in a subdirectory D:/s. > This may have something to do with it... or may be extraneous. > > You haven't indicated what your working environment is, though the OS> mention a distinction between whatever this environment is (RGui?) and > "console". Are you using Cygwin? could end-of-line termination (CRLF vs > LF) be causing you difficulty?The OP explained that and if you believe OP changing intended file changes the error message. And, yes, I would strongly suggest getting cygwin so you have some tools other than posting incomplete information of calling tech support LOL. In this case, you would use something like "od" to verify that your file is as you expect. Just as printing numerical data often truncated digits, unprintable chars don't always show up on printing. od -ax or something may be informative. Changing the tab may have caused editor to change line endings or something else. Smart editors often mess stuff up.> > Perhaps you should follow the posting guide instructions... > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 2 Date: Mon, 22 Nov 2010 03:36:17 -0800 (PST) From: madr <madrazel at interia.pl> To: r-help at r-project.org Subject: [R] how to sample lowess/loess into matrix ? Message-ID: <1290425777216-3053458.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii code: x <- rnorm(32) y <- rnorm(32) plot(x,y) lines(lowess(x,y),col='red') Now I need to sample the lowess function into matrix where one series will be X and other will be values of lowess at particular X. -- View this message in context: http://r.789695.n4.nabble.com/how-to-sample-lowess-loess-into-matrix-tp30534 58p3053458.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 3 Date: Mon, 22 Nov 2010 06:38:52 -0500 From: Duncan Murdoch <murdoch.duncan at gmail.com> To: madr <madrazel at interia.pl> Cc: r-help at r-project.org Subject: Re: [R] "unexpected numeric constant" while reading tab delimited csv file Message-ID: <4CEA564C.4000908 at gmail.com> Content-Type: text/plain; charset=ISO-8859-1; format=flowed madr wrote:> my csv file is very simple - just one line for purpose of this test: > 0{TAB}0 > > and read function is this: > csvdata = read.csv(file="d:/s/test.csv",head=FALSE,sep="\t") > > then error comes: > > Error in source("d:/test.csv") : > d:/test.csv:1:9: unexpected numeric constant > 1: 0 0 > > > but when I change delimiter to ; (colon) then error not shows up anymoreYou used source, not read.csv. They aren't the same thing. If you typed what you said you typed, then you've hidden the real read.csv function behind your own, and your own calls source. But I don't think you typed what you said you typed. Duncan Murdoch ------------------------------ Message: 4 Date: Mon, 22 Nov 2010 20:11:39 +0800 From: Luis Felipe Parra <felipe.parra at quantil.com.co> To: r-help <r-help at r-project.org> Subject: [R] Rexcel Message-ID: <AANLkTi=rWSJheF2WLtmhe8Lg0byHZf1nr3Cdj18Q8+mq at mail.gmail.com> Content-Type: text/plain Hello I am new to RExcel and I would like to run a source code form the excel worksheet. I would like to run the following code source("C:\\Quantil Aplicativos\\Genercauca\\BackwardSelectionNC.r") from the excel wroksheet. Does anybody know how to do this? Thank you Felipe Parra [[alternative HTML version deleted]] ------------------------------ Message: 5 Date: Mon, 22 Nov 2010 23:22:25 +1100 From: Nevil Amos <nevil.amos at gmail.com> To: R-help at r-project.org Subject: [R] how to change number of characters per line for print() to sink()? Message-ID: <4CEA6081.70202 at sci.monash.edu.au> Content-Type: text/plain; charset=ISO-8859-1; format=flowed I am using r to read and reformat data that is then saved to a text file using sink(), the file has a number of initial lines of comments and summary data followed by print of a data.frame with no row names. for example a<-c(100:120) b<-c(rnorm(100:120)) c<-c(rnorm(200:220)) mydata<-data.frame(rbind(a,b,c)) sink("datafile.txt") cat("comments about my data \n") cat("other calculations returned as separate text comments on a line \n") print(mydata,row.names=F) sink() I need the content of the text file to keep each row of the data frame on a single line thus (with intervening columns present of course) "datafile.txt" comments about my data other calculations returned as separate text comments on a line X1 X2 X3 X4 X5 X6 ..................................................................... X19 X20 X21 100.0000000 101.000000 102.0000000 103.0000000 104.0000000 105.0000000 ......................118.0000000 119.0000000 120.0000000 -0.3380570 -1.400905 1.0396499 -0.5802181 -0.2340614 0.6044928 ...................................-0.4854702 -0.3677461 -1.2033173 -0.9002824 1.544242 -0.8668653 0.3066256 0.2490254 -1.6429223 ..................................... 0.0861146 0.4276929 -0.3408604 How doI change setting for print() or use another function to keep each row of the data frame as a single line ( of greater length up to approx 300 characters) instead of wrapping the data frame into multiple lines of text? The problem : I end up with the data frame split into several sections one under another thus "datafile.txt" comments about my data other calculations returned as separate text comments on a line X1 X2 X3 X4 X5 X6 100.0000000 101.000000 102.0000000 103.0000000 104.0000000 105.0000000 -0.3380570 -1.400905 1.0396499 -0.5802181 -0.2340614 0.6044928 -0.9002824 1.544242 -0.8668653 0.3066256 0.2490254 -1.6429223 X7 X8 X9 X10 X11 X12 106.0000000 107.00000000 108.0000000 109.0000000 110.0000000 111.0000000 0.3152427 0.15093494 -0.3316172 -0.3603724 -2.0516402 -0.4556241 -0.6502265 -0.08842649 -0.3775335 -0.4942572 -0.0976565 -0.7716651 X13 X14 X15 X16 X17 X18 112.0000000 113.0000000 114.0000000 115.0000000 116.00000000 117.000000 0.8829135 0.8851043 -0.7687383 -0.9573476 -0.03041968 1.425754 0.2666777 0.6405255 0.2342905 -0.7705545 -1.18028004 1.303601 X19 X20 X21 118.0000000 119.0000000 120.0000000 -0.4854702 -0.3677461 -1.2033173 0.0861146 0.4276929 -0.3408604 ------------------------------ Message: 6 Date: Mon, 22 Nov 2010 04:31:29 -0800 (PST) From: Aline Uwimana via LinkedIn <member at linkedin.com> To: James Holtman <r-help at r-project.org> Subject: [R] Invitation ? se connecter sur LinkedIn Message-ID: <1473926269.6681467.1290429089742.JavaMail.app at ech3-cdn12.prod> Content-Type: text/plain LinkedIn ------------Aline Uwimana requested to add you as a connection on LinkedIn: ------------------------------------------ James, J'aimerais vous inviter C rejoindre mon rC)seau professionnel en ligne, sur le site LinkedIn. Aline Accept invitation from Aline Uwimana http://www.linkedin.com/e/j2w180-ggtc5u4r-5e/dwA6EuZvdiRUO9LjukuviNUAHj8vHaN g/blk/I2473001097_2/1BpC5vrmRLoRZcjkkZt5YCpnlOt3RApnhMpmdzgmhxrSNBszYOnPsVc3 4Mc3cTd399bSlUk6ZDtQd6bPkOcP4Sd3kPej4LrCBxbOYWrSlI/EML_comm_afe/ View invitation from Aline Uwimana http://www.linkedin.com/e/j2w180-ggtc5u4r-5e/dwA6EuZvdiRUO9LjukuviNUAHj8vHaN g/blk/I2473001097_2/39vdPAMcj0McPsQcAALqnpPbOYWrSlI/svi/ ------------------------------------------ DID YOU KNOW you can be the first to know when a trusted member of your network changes jobs? With Network Updates on your LinkedIn home page, you'll be notified as members of your network change their current position. Be the first to know and reach out! http://www.linkedin.com/ -- (c) 2010, LinkedIn Corporation [[alternative HTML version deleted]] ------------------------------ Message: 7 Date: Mon, 22 Nov 2010 07:32:11 -0500 From: Gabor Grothendieck <ggrothendieck at gmail.com> To: rnick <nikos.rachmanis at gmail.com> Cc: r-help at r-project.org Subject: Re: [R] How to combine Date and time in one column Message-ID: <AANLkTimU-LvRyziVgf0n_=W+WTH0TPYqgR4ZvxkWSXSu at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 On Mon, Nov 22, 2010 at 2:24 AM, rnick <nikos.rachmanis at gmail.com> wrote:> > Hello everyone, > > I am trying to built an xts object and i have run into some problems onthe> data handling. I would really appreciate if someone could help me with the > following: > > 1) I have a OHLC dataset with Time and date in different columns. Howcould> i combine date and time in one column in order to pass on the new columnto> xts? I have use cbind and data.frame before but i did not manage to yield > any good results as the formating of the file changes. > > Date ? ? ? ? ? ?Time ? ? ? ? ? O ? ? ? ? ? ? ? ?H ? ? ? ? ? ? ? L ? ? ? ?? ? ? C> 1/2/2005 ? ? ? ?17:05 ? ? ? ? ?1.3546 ? 1.3553 ?1.3546 ?1.35495 > 1/2/2005 ? ? ? ?17:10 ? ? ? ? ?1.3553 ? 1.3556 ?1.3549 ?1.35525 > 1/2/2005 ? ? ? ?17:15 ? ? ? ? ?1.3556 ? 1.35565 1.35515 1.3553 > 1/2/2005 ? ? ? ?17:25 ? ? ? ? ?1.355 ? ? ? ? ? ?1.3556 ?1.355 ? ? ? ? ? 1.3555> 1/2/2005 ? ? ? ?17:30 ? ? ? ? ?1.3556 ? 1.3564 ?1.35535 1.3563 > > 2) It is not clear to me what is the best way to construct the .xtsobject?> Should i use only the Date&time to index or should i also combine it with > the rest of the variables? >Use read.zoo and then as.xts to convert it to xts. The following shows it for chron date/times. Replace textConnection(Lines) with "myfile.dat" to read it from that file. You can replace the FUNpart with a conversion to any date/time class supported by xts. Here we show it for chron. In the example below we are assuming that the date format is month/day/year. See R News 4/1. Lines <- "Date Time O H L C 1/2/2005 17:05 1.3546 1.3553 1.3546 1.35495 1/2/2005 17:10 1.3553 1.3556 1.3549 1.35525 1/2/2005 17:15 1.3556 1.35565 1.35515 1.3553 1/2/2005 17:25 1.355 1.3556 1.355 1.3555 1/2/2005 17:30 1.3556 1.3564 1.35535 1.3563" library(xts) # this also pulls in zoo and its read.zoo library(chron) z <- read.zoo(textConnection(Lines), header = TRUE, index = list(1, 2), FUN = function(d, t) as.chron(paste(as.Date(chron(d)), t))) x <- as.xts(z) -- Statistics & Software Consulting GKX Group, GKX Associates Inc. tel: 1-877-GKX-GROUP email: ggrothendieck at gmail.com ------------------------------ Message: 8 Date: Mon, 22 Nov 2010 09:43:18 -0300 From: Jose Claudio Faria <joseclaudio.faria at gmail.com> To: r-help at r-project.org Subject: [R] Tinn-R 2.3.7.0 released Message-ID: <AANLkTimGQHOWj_EzDhR_J-38QxG99KDGmq0OvGZ29eiZ at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Dears users, A new version of Tinn-R was released today. Below details: 2.3.7.0 (Nov/22/2010) * Bug(s) fixed: - A bug related with the intermittent loose of connection (or appear to freeze) with Rgui.exe was fixed. * The versions 2.3.6.4, 2.3.6.5, 2.3.6.6 and 2.3.6.7 restricted to pre-release testers. * The Application options interface was a bit changed: - The Application options/R/Rterm was split in two tabs: Error and Options. The tab Error has a new option: Try to find error in the editor. It enables the user to set Tinn-R in order to find errors in the editor when sending instructions to Rterm. * This version is full compatible with Windows 7 and R 2.12.0. * The component XPmenu was removed from the project. Windows XP users, perhaps, will find the Tinn-R appearance less attractive, but the applicative is now more stable. As soon as possible, the project will get a better option for skins. * Parts of the source code were optimized. All the best, -- ///\\\///\\\///\\\///\\\///\\\///\\\///\\\///\\\ Jose Claudio Faria Estatistica - prof. Titular UESC/DCET/Brasil joseclaudio.faria at gmail.com ///\\\///\\\///\\\///\\\///\\\///\\\///\\\///\\\ ------------------------------ Message: 9 Date: Mon, 22 Nov 2010 20:53:14 +0800 From: Luis Felipe Parra <felipe.parra at quantil.com.co> To: r-help <r-help at r-project.org> Subject: [R] Rexcel Message-ID: <AANLkTimtvxE9BvWqZ+=bK7LZOcAou_jrA3OFDJOZ5Gcj at mail.gmail.com> Content-Type: text/plain Hello, I am trying to use RExcel and I would like to know if it is possible to use in excel the following function I made for R Pron = function(path="C:\\Quantil Aplicativos\\Genercauca\\V5\\"){ library(timeSeries) library(maSigPro) ### CARGAR FUNCIONES source(paste(path,"\\fUtilidades\\BackwardSelectionNC.r",sep<file://futilida des//BackwardSelectionNC.r%22,sep> ="")) source(paste(path,"\\fUtilidades\\CriteriosDeComparacion.r",sep<file://futil idades//CriteriosDeComparacion.r%22,sep> ="")) dataTSORG<-read.csv('entrada.csv', header = TRUE, sep = ",", quote="\"", dec=".",fill = TRUE, comment.char="") dataTSORG = ts(dataTSORG, start=c(1950,1), frequency=12) dataTSORG = as.timeSeries(dataTSORG) X = prcomp(dataTSORG[,2:40])$x return(X) } Does somebody know if its possible? and if so how can I do it? Thank you Felipe Parra [[alternative HTML version deleted]] ------------------------------ Message: 10 Date: Mon, 22 Nov 2010 07:59:05 -0500 From: jim holtman <jholtman at gmail.com> To: nevil.amos at sci.monash.edu.au Cc: R-help at r-project.org Subject: Re: [R] how to change number of characters per line for print() to sink()? Message-ID: <AANLkTi=_J6jSVAuVX8SiF8g+9w6pVLPx7v+HTqa-4_wm at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 ?options width options(width = 1000) On Mon, Nov 22, 2010 at 7:22 AM, Nevil Amos <nevil.amos at gmail.com> wrote:> I am using r to read and reformat data that is then saved to a text file > using sink(), the file has a number of initial lines of comments andsummary> data followed by print of a data.frame with no row names. > for example > > a<-c(100:120) > b<-c(rnorm(100:120)) > c<-c(rnorm(200:220)) > mydata<-data.frame(rbind(a,b,c)) > sink("datafile.txt") > > cat("comments about my data \n") > cat("other calculations returned as separate text comments on a line \n") > print(mydata,row.names=F) > sink() > > > I need the content of the text file to keep each row of the data frame ona> single line thus (with intervening columns present of course) > > "datafile.txt" > > comments about my data > other calculations returned as separate text comments on a line > ? ? ? ? ?X1 ? ? ? ? X2 ? ? ? ? ?X3 ? ? ? ? ?X4 ? ? ? ? ?X5 ? ? ? ? ?X6 > ? ? ..................................................................... > ? ? ? X19 ? ? ? ? X20 ? ? ? ? X21 > > > ?100.0000000 101.000000 102.0000000 103.0000000 104.0000000 105.0000000 > ?......................118.0000000 119.0000000 120.0000000 > ?-0.3380570 ?-1.400905 ? 1.0396499 ?-0.5802181 ?-0.2340614 ? 0.6044928 > ...................................-0.4854702 ?-0.3677461 ?-1.2033173 > ?-0.9002824 ? 1.544242 ?-0.8668653 ? 0.3066256 ? 0.2490254 ?-1.6429223 > ..................................... ? 0.0861146 ? 0.4276929 ?-0.3408604 > > How doI change setting for print() or use another function to keep eachrow> of the data frame as a single line ( of greater length up to approx 300 > characters) instead of wrapping the data frame into multiple lines oftext?> > The problem : I end up with the data frame split into several sections one > under another thus > > "datafile.txt" > > comments about my data > other calculations returned as separate text comments on a line > ? ? ? ? ?X1 ? ? ? ? X2 ? ? ? ? ?X3 ? ? ? ? ?X4 ? ? ? ? ?X5 ? ? ? ? ?X6 > ?100.0000000 101.000000 102.0000000 103.0000000 104.0000000 105.0000000 > ?-0.3380570 ?-1.400905 ? 1.0396499 ?-0.5802181 ?-0.2340614 ? 0.6044928 > ?-0.9002824 ? 1.544242 ?-0.8668653 ? 0.3066256 ? 0.2490254 ?-1.6429223 > ? ? ? ? ?X7 ? ? ? ? ? X8 ? ? ? ? ?X9 ? ? ? ? X10 ? ? ? ? X11 ? ? ? ? X12 > ?106.0000000 107.00000000 108.0000000 109.0000000 110.0000000 111.0000000 > ? 0.3152427 ? 0.15093494 ?-0.3316172 ?-0.3603724 ?-2.0516402 ?-0.4556241 > ?-0.6502265 ?-0.08842649 ?-0.3775335 ?-0.4942572 ?-0.0976565 ?-0.7716651 > ? ? ? ? X13 ? ? ? ? X14 ? ? ? ? X15 ? ? ? ? X16 ? ? ? ? ?X17 ? ? ? ?X18 > ?112.0000000 113.0000000 114.0000000 115.0000000 116.00000000 117.000000 > ? 0.8829135 ? 0.8851043 ?-0.7687383 ?-0.9573476 ?-0.03041968 ? 1.425754 > ? 0.2666777 ? 0.6405255 ? 0.2342905 ?-0.7705545 ?-1.18028004 ? 1.303601 > ? ? ? ? X19 ? ? ? ? X20 ? ? ? ? X21 > ?118.0000000 119.0000000 120.0000000 > ?-0.4854702 ?-0.3677461 ?-1.2033173 > ? 0.0861146 ? 0.4276929 ?-0.3408604 > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >-- Jim Holtman Cincinnati, OH +1 513 646 9390 What is the problem that you are trying to solve? ------------------------------ Message: 11 Date: Mon, 22 Nov 2010 13:02:17 +0000 From: "S Ellison" <S.Ellison at lgc.co.uk> To: "emorway" <emorway at engr.colostate.edu>, "Uwe Ligges" <ligges at statistik.tu-dortmund.de> Cc: r-help at r-project.org Subject: Re: [R] boxplot: reverse y-axis order Message-ID: <scea69e8.069 at tedmail.lgc.co.uk> Content-Type: text/plain; charset=US-ASCII A simple alternative is to use "at" to contrl plot locations: boxplot( ..., at=rev(1:nlevels(depthM))) which just rearranges where they are plotted. Example: set.seed(1023) x <- gl(5, 5) y<-rnorm(25) boxplot(y~x, horizontal=TRUE) boxplot(y~x, at=rev(1:nlevels(x)), , horizontal=TRUE) Steve E>>> Uwe Ligges <ligges at statistik.tu-dortmund.de> 21/11/2010 19:38:54 >>>On 21.11.2010 20:30, emorway wrote:> > Hello, > > Searching this forum has enabled me to get pretty far in what I'mtrying to> do. However, there is one more manipulation I would like to make andI> haven't found a solution. Using the data and code below, I generatethe> plot produced by the last command. If possible I would like toreverse the> order of the y-axis (bearing in mind horizontal=T) so that 0 isplotted at> the upper most part of the y-axis and 3 at the axis intercepts. I'vetried> the experiment approach to no avail, meaning I've placed rev(...)around> various arguments but with wrong results. > > Thanks, > Eric > > > df<-read.table(textConnection("Field Date AvgWTD Region variablevalue hole> depth depthM > 204 17-Aug-00 2.897428989 US R1 NA R 1 0[SNIP]> 9A 09-Aug-00 1.482089996 US C6 NA C 6 1.8 > 9B 01-Jun-01 1.409700036 US C6 NA C 6 1.8 > 9B 09-Aug-00 3.837660074 US C6 NA C 6 1.8"),header=T) > closeAllConnections() > #The folling call doesn't preserve the correct spacing between data > boxplot(value~depthM,data=df,horizontal=T,outline=F) > #The following command preserves correct spacing, but need to reversethe> order >boxplot(value~factor(depthM,levels=c(0.0,0.3,0.6,0.9,1.2,1.5,1.8,2.1,2.4,2.7 ,3)),data=df,horizontal=T,outline=F) So if you want to reverse, either specify the levels in reverse order or use rev() as in: boxplot(value ~ factor(depthM, levels = rev(c(0.0,0.3,0.6,0.9,1.2,1.5,1.8,2.1,2.4,2.7,3))), data = df, horizontal = TRUE, outline = FALSE) Uwe Ligges ______________________________________________ R-help at r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. ******************************************************************* This email and any attachments are confidential. Any use...{{dropped:8}} ------------------------------ Message: 12 Date: Mon, 22 Nov 2010 05:19:21 -0800 (PST) From: Alaios <alaios at yahoo.com> To: Rhelp <r-help at r-project.org> Subject: [R] plot inside function does not work Message-ID: <413924.42773.qm at web120105.mail.ne1.yahoo.com> Content-Type: text/plain Hello everyone, when I commit a plot using console(command line?) plot works fine. I have created a function that plots based on the input. This function is called plot_shad. When I call this function alone in the command line I get my plot. Then I tried to use another function as depicted_below to do some calculation before calling the function that does the plotting. plot_shad_map<-function(f,CRagent,agentid){ for (i in c(1:nrow(shad_map))){ for (j in c(1:ncol(shad_map))){ # Do something } } plot_shad_f(shad_map) # This plots fine when used in command line. But inside this #function does not return(shad_map) } Unfortunately I get no plot . What might be the problem? One more question how to get more plots at the same time. It seems that when I issue a new plot replaces the old plot. I would like to thank you in advance for you help Regards Alex [[alternative HTML version deleted]] ------------------------------ Message: 13 Date: Mon, 22 Nov 2010 05:32:05 -0800 (PST) From: madr <madrazel at interia.pl> To: r-help at r-project.org Subject: Re: [R] how to sample lowess/loess into matrix ? Message-ID: <1290432725271-3053601.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii I found it, it was SO simple: lowessline <- lowess(x,y) write.csv(lowessline, "loess.csv") -- View this message in context: http://r.789695.n4.nabble.com/how-to-sample-lowess-loess-into-matrix-tp30534 58p3053601.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 14 Date: Mon, 22 Nov 2010 14:34:02 +0100 From: Claudia Beleites <cbeleites at units.it> To: r-help at r-project.org Subject: Re: [R] plot inside function does not work Message-ID: <4CEA714A.5070406 at units.it> Content-Type: text/plain; charset=UTF-8; format=flowed Alex, this may be FAQ 7.22 Claudia On 11/22/2010 02:19 PM, Alaios wrote:> Hello everyone, > when I commit a plot using console(command line?) plot works fine. I havecreated a function that plots based on the input. This function is called plot_shad. When I call this function alone in the command line I get my plot.> > Then I tried to use another function as depicted_below to do somecalculation before calling the function that does the plotting.> > plot_shad_map<-function(f,CRagent,agentid){ > ? for (i in c(1:nrow(shad_map))){ > ??? for (j in c(1:ncol(shad_map))){ > ???? # Do something > ??? } > ? } > ? plot_shad_f(shad_map) # This plots fine when used in command line. Butinside this #function does not> ? return(shad_map) > } > > Unfortunately I get no plot . What might be the problem? > > One more question how to get more plots at the same time. It seems thatwhen I issue a new plot replaces the old plot.> > I would like to thank you in advance for you help > Regards > Alex > > > > > [[alternative HTML version deleted]] > > > > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.-- Claudia Beleites Dipartimento dei Materiali e delle Risorse Naturali Universit? degli Studi di Trieste Via Alfonso Valerio 6/a I-34127 Trieste phone: +39 0 40 5 58-37 68 email: cbeleites at units.it ------------------------------ Message: 15 Date: Mon, 22 Nov 2010 05:36:06 -0800 (PST) To: Luis Felipe Parra <felipe.parra at quantil.com.co>, r-help <r-help at r-project.org> Subject: Re: [R] Rexcel Message-ID: <666026.40186.qm at web113206.mail.gq1.yahoo.com> Content-Type: text/plain; charset=utf-8 Hi, For RExcel I would suggest subscribing; http://mailman.csd.univie.ac.at/listinfo/rcom-l They have a website on; http://rcom.univie.ac.at/ B.R. Stephen L ----- Original Message ---- From: Luis Felipe Parra <felipe.parra at quantil.com.co> To: r-help <r-help at r-project.org> Sent: Mon, November 22, 2010 8:11:39 PM Subject: [R] Rexcel Hello I am new to RExcel and I would like to run a source code form the excel worksheet. I would like to run the following code source("C:\\Quantil Aplicativos\\Genercauca\\BackwardSelectionNC.r") from the excel wroksheet. Does anybody know how to do this? Thank you Felipe Parra [[alternative HTML version deleted]] ______________________________________________ R-help at r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. ------------------------------ Message: 16 Date: Mon, 22 Nov 2010 05:47:35 -0800 (PST) To: r-help at r-project.org, Claudia Beleites <cbeleites at units.it> Subject: Re: [R] plot inside function does not work Message-ID: <946840.29824.qm at web120104.mail.ne1.yahoo.com> Content-Type: text/plain Dear Claudia, I would like to thank you for your reply, according to 7.22 I have to put some print() statement inside my function. What I do not understand is where to put this line (at the beginning or at the end of the code?) I tried both but I get error message. Moreover I would like to ask you if any inputs should be passed to the print(). I tried to pass inside the argument of ggplot but it didnt like it :( Best Regards 7.22 Why do lattice/trellis graphics not work? The most likely reason is that you forgot to tell R to display the graph. Lattice functions such as xyplot() create a graph object, but do not display it (the same is true of ggplot2 graphics, and Trellis graphics in S-Plus). The print() method for the graph object produces the actual display. When you use these functions interactively at the command line, the result is automatically printed, but in source() or inside your own functions you will need an explicit print() statement. --- On Mon, 11/22/10, Claudia Beleites <cbeleites at units.it> wrote: From: Claudia Beleites <cbeleites at units.it> Subject: Re: [R] plot inside function does not work To: r-help at r-project.org Date: Monday, November 22, 2010, 1:34 PM Alex, this may be FAQ 7.22 Claudia On 11/22/2010 02:19 PM, Alaios wrote:> Hello everyone, > when I commit a plot using console(command line?) plot works fine. I havecreated a function that plots based on the input. This function is called plot_shad. When I call this function alone in the command line I get my plot.> > Then I tried to use another function as depicted_below to do somecalculation before calling the function that does the plotting.> > plot_shad_map<-function(f,CRagent,agentid){ > o?= for (i in c(1:nrow(shad_map))){ > o?=o?=o?= for (j in c(1:ncol(shad_map))){ > o?=o?=o?=o?= # Do something > o?=o?=o?= } > o?= } > o?= plot_shad_f(shad_map) # This plots fine when used in command line. Butinside this #function does not> o?= return(shad_map) > } > > Unfortunately I get no plot . What might be the problem? > > One more question how to get more plots at the same time. It seems thatwhen I issue a new plot replaces the old plot.> > I would like to thank you in advance for you help > Regards > Alex > > > > > B B B [[alternative HTML version deleted]] > > > > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.-- Claudia Beleites Dipartimento dei Materiali e delle Risorse Naturali UniversitC degli Studi di Trieste Via Alfonso Valerio 6/a I-34127 Trieste phone: +39 0 40 5 58-37 68 email: cbeleites at units.it ______________________________________________ R-help at r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. [[alternative HTML version deleted]] ------------------------------ Message: 17 Date: Mon, 22 Nov 2010 05:53:40 -0800 (PST) From: Yann Lancien via LinkedIn <member at linkedin.com> To: James Holtman <r-help at r-project.org> Subject: [R] Invitation ? se connecter sur LinkedIn Message-ID: <210238489.6957030.1290434020449.JavaMail.app at ech3-cdn05.prod> Content-Type: text/plain LinkedIn ------------Yann Lancien requested to add you as a connection on LinkedIn: ------------------------------------------ James, J'aimerais vous inviter C rejoindre mon rC)seau professionnel en ligne, sur le site LinkedIn. Yann Accept invitation from Yann Lancien http://www.linkedin.com/e/j2w180-ggtf3iot-z/dwA6EuZvdiRUO9LjukuviNUAHj8vHaNg /blk/I2473133047_2/1BpC5vrmRLoRZcjkkZt5YCpnlOt3RApnhMpmdzgmhxrSNBszYOnPsQc3c PcjcTd399bPxgum8Utk5QbPcNdzkQe3kPej4LrCBxbOYWrSlI/EML_comm_afe/ View invitation from Yann Lancien http://www.linkedin.com/e/j2w180-ggtf3iot-z/dwA6EuZvdiRUO9LjukuviNUAHj8vHaNg /blk/I2473133047_2/39vdPgMcPcNcPsQcAALqnpPbOYWrSlI/svi/ ------------------------------------------ Why might connecting with Yann Lancien be a good idea? People Yann Lancien knows can discover your profile: Connecting to Yann Lancien will attract the attention of LinkedIn users. See who's been viewing your profile: http://www.linkedin.com/e/j2w180-ggtf3iot-z/wvp/inv18_wvmp/ -- (c) 2010, LinkedIn Corporation [[alternative HTML version deleted]] ------------------------------ Message: 18 Date: Mon, 22 Nov 2010 06:06:30 -0800 (PST) From: Joel <joda2457 at student.uu.se> To: r-help at r-project.org Subject: [R] Prob with merge Message-ID: <1290434790326-3053652.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii Hi Im trying to merge 2 data frames using merge but I dont get the result i want Lets make this a small test as my data set is to big to put in here :). t1<-data.frame(a=c(1,2,3,4,5,6),b=c(11,11,NA,11,11,11)) t1<-data.frame(a=c(1,2,3,4,5,8),b=c(12,12,12,12,12,32)) this gives me:> t1a b 1 1 12 2 2 12 3 3 12 4 4 12 5 5 12 6 8 32> t2a b 1 1 11 2 2 11 3 3 NA 4 4 11 5 5 11 6 6 11 now when i merge i get:> merge(t1,t2, by="a")a b.x b.y 1 1 12 11 2 2 12 11 3 3 12 NA 4 4 12 11 5 5 12 11 But what I want is it to look like: a b.x b.y 1 1 12 11 2 2 12 11 3 3 12 NA 4 4 12 11 5 5 12 11 6 8 32 NA So I keep all of the rows from t1 and get an NA in dose slots at the t2 part of the merge. Anyone know how to accomplice this? Thx //Joel -- View this message in context: http://r.789695.n4.nabble.com/Prob-with-merge-tp3053652p3053652.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 19 Date: Mon, 22 Nov 2010 06:08:10 -0800 From: Peter Ehlers <ehlers at ucalgary.ca> To: Luis Felipe Parra <felipe.parra at quantil.com.co> Cc: r-help <r-help at r-project.org> Subject: Re: [R] Making a line in a legend shorter Message-ID: <4CEA794A.3000607 at ucalgary.ca> Content-Type: text/plain; charset=ISO-8859-1; format=flowed On 2010-11-21 18:38, Luis Felipe Parra wrote:> Hello, I am putting a legend with lines in a line plot and I would like to > make the lines in the legend shorter. Does anybody knows how to do this?I think the segment length is still hard-coded in legend(). Last July there was a request to lengthen the segment. Maybe this post will help you: https://stat.ethz.ch/pipermail/r-help/2010-July/246599.html Peter Ehlers> > Thank you > > Felipe Parra > > [[alternative HTML version deleted]]------------------------------ Message: 20 Date: Mon, 22 Nov 2010 15:09:31 +0100 From: "ONKELINX, Thierry" <Thierry.ONKELINX at inbo.be> To: Joel <joda2457 at student.uu.se>, <r-help at r-project.org> Subject: Re: [R] Prob with merge Message-ID: <3DB16098F738284D8DBEB2FC369916384D6283 at inboexch.inbo.be> Content-Type: text/plain; charset="us-ascii" merge(t1,t2, by="a", all.x = TRUE) ------------------------------------------------------------------------ ---- ir. Thierry Onkelinx Instituut voor natuur- en bosonderzoek team Biometrie & Kwaliteitszorg Gaverstraat 4 9500 Geraardsbergen Belgium Research Institute for Nature and Forest team Biometrics & Quality Assurance Gaverstraat 4 9500 Geraardsbergen Belgium tel. + 32 54/436 185 Thierry.Onkelinx at inbo.be www.inbo.be To call in the statistician after the experiment is done may be no more than asking him to perform a post-mortem examination: he may be able to say what the experiment died of. ~ Sir Ronald Aylmer Fisher The plural of anecdote is not data. ~ Roger Brinner The combination of some data and an aching desire for an answer does not ensure that a reasonable answer can be extracted from a given body of data. ~ John Tukey> -----Oorspronkelijk bericht----- > Van: r-help-bounces at r-project.org > [mailto:r-help-bounces at r-project.org] Namens Joel > Verzonden: maandag 22 november 2010 15:07 > Aan: r-help at r-project.org > Onderwerp: [R] Prob with merge > > > Hi > > Im trying to merge 2 data frames using merge but I dont get > the result i want > > Lets make this a small test as my data set is to big to put > in here :). > > t1<-data.frame(a=c(1,2,3,4,5,6),b=c(11,11,NA,11,11,11)) > t1<-data.frame(a=c(1,2,3,4,5,8),b=c(12,12,12,12,12,32)) > > this gives me: > > > t1 > a b > 1 1 12 > 2 2 12 > 3 3 12 > 4 4 12 > 5 5 12 > 6 8 32 > > t2 > a b > 1 1 11 > 2 2 11 > 3 3 NA > 4 4 11 > 5 5 11 > 6 6 11 > > now when i merge i get: > > merge(t1,t2, by="a") > a b.x b.y > 1 1 12 11 > 2 2 12 11 > 3 3 12 NA > 4 4 12 11 > 5 5 12 11 > > > But what I want is it to look like: > > a b.x b.y > 1 1 12 11 > 2 2 12 11 > 3 3 12 NA > 4 4 12 11 > 5 5 12 11 > 6 8 32 NA > > So I keep all of the rows from t1 and get an NA in dose slots > at the t2 part of the merge. > Anyone know how to accomplice this? > > Thx > //Joel > > > > -- > View this message in context: > http://r.789695.n4.nabble.com/Prob-with-merge-tp3053652p3053652.html > Sent from the R help mailing list archive at Nabble.com. > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. >------------------------------ Message: 21 Date: Mon, 22 Nov 2010 08:11:15 -0600 From: Terry Therneau <therneau at mayo.edu> Cc: r-help at r-project.org, dieter.menne at menne-biomed.de, r_tingley at hotmail.com Subject: Re: [R] calculating martingale residual on new data using "predict.coxph" Message-ID: <1290435075.7350.3.camel at punchbuggy> Content-Type: text/plain This feature has been added in survival 2.36-1, which is now on CRAN. (2.36-2 should appear in another day or so) Terry T. ---------begin included message -------- I was trying to use "predict.coxph" to calculate martingale residuals on a test data, however, as pointed out before http://tolstoy.newcastle.edu.au/R/e4/help/08/06/13508.html predict(mycox1, newdata, type="expected") is not implemented yet. ------------------------------ Message: 22 Date: Mon, 22 Nov 2010 06:20:51 -0800 (PST) From: Joel <joda2457 at student.uu.se> To: r-help at r-project.org Subject: Re: [R] Prob with merge Message-ID: <1290435651761-3053675.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii Thx alot mate. -- View this message in context: http://r.789695.n4.nabble.com/Prob-with-merge-tp3053652p3053675.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 23 Date: Mon, 22 Nov 2010 09:27:08 -0500 From: Jimmy Martina <jimmyjuan1 at hotmail.com> To: R-geo <r-sig-geo at stat.math.ethz.ch>, R <r-help at r-project.org> Subject: [R] What if geoRglm results showed that a non-spacial model fits? Message-ID: <BAY110-W26A01692A4E820BEE2C8BF943D0 at phx.gbl> Content-Type: text/plain Hi R-people: Working in geoRglm, it shows me, according to AIC criterion, that the non-spacial model describes the process in a better way. It's the first time that I'm facing up to. These are my results: OP2003Seppos.AICnonsp-OP2003Seppos.AICsp #[1] -4 (OP2003Seppos.lf0.p<-exp(OP2003Seppos.lf0$beta)/(1+exp(OP2003Seppos.lf0$beta ))) #P non spatial #[1] 0.9717596 (OP2003Seppos.lf.p<-exp(OP2003Seppos.lf$beta)/(1+exp(OP2003Seppos.lf$beta))) #P spatial #[1] 0.9717596 It must what have an important influence at kriging, because it shows as following: OP2003Sepposbin.krig<-glsm.krige(OP2003Seppos.tune,loc=OP2003Seppospro.pred. grid,bor=OP2003Sepposbor) #glsm.krige: Prediction for a generalised linear spatial model #There are 50 or mode advices (use warnings() to see them) #> warnings() #Warning messages: #1: In asympvar(kpl.result$predict, messages = FALSE) ... : # value of argument lag.max is not suffiently long #2: In asympvar(kpl.result$predict, messages = FALSE) ... : # value of argument lag.max is not suffiently long Help me, please. [[alternative HTML version deleted]] ------------------------------ Message: 24 Date: Mon, 22 Nov 2010 15:32:51 +0100 From: Stella Pachidi <stella.pachidi at gmail.com> To: r-help at stat.math.ethz.ch Subject: [R] Help with plotting kohonen maps Message-ID: <AANLkTini0qSwDHVo_Pe9f2zh2Ga5mEWfBWhCXiwwQWYr at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Dear all, I recently started using the kohonen package for my thesis project. I have a very simple question which I cannot figure out by myself: When I execute the following example code, from the paper of Wehrens and Buydens (http://www.jstatsoft.org/v21/i05/paper): R> library("kohonen") Loading required package: class R> data("wines") R> wines.sc <- scale(wines) R> set.seed(7) R> wine.som <- som(data = wines.sc, grid = somgrid(5, 4, "hexagonal")) R> plot(wine.som, main = "Wine data") I get to have a plot of the codebook vectors of the 5-by-4 mapping of the wine data, and it also includes which variable names correspond to each color. (same picture as in the paper) However, when I run the som() function with my own data and I try to get the plot afterwards: library("kohonen") self_Organising_Map <- som(data = tableToCluster, grid = somgrid(5, 2, "rectangular"), rlen=1000) plot(self_Organising_Map, main = "Kohonen Map of Clustered Profiles") the resulting plot does not contain the color labels i.e. the variable names of my data table, even though they exist and are included as column names of tableToCluster. I also tried the following line: plot(self_Organising_Map, type="codes", codeRendering = "segments", ncolors=length(colnames(self_Organising_Map$codes)), palette.name=rainbow, main = "Kohonen Map of Clustered Profiles \n Codes", zlim =colnames(self_Organising_Map$codes)) but it had the same result. If you could please help with what argument I should use to show the [[elided Yahoo spam]] Kind regards, Stella -- Stella Pachidi Master in Business Informatics student Utrecht University ------------------------------ Message: 25 Date: Mon, 22 Nov 2010 12:03:54 -0300 From: Kjetil Halvorsen <kjetilbrinchmannhalvorsen at gmail.com> To: Uwe Ligges <ligges at statistik.tu-dortmund.de> Cc: r-devel <r-help at r-project.org> Subject: Re: [R] ?summaryRprof running at 100% cpu for one hour ... Message-ID: <AANLkTim7qsgaidiEV65pbJfPLOhrHG4yy8jf6Eha==t9 at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 see below. 2010/11/21 Uwe Ligges <ligges at statistik.tu-dortmund.de>:> > > On 21.11.2010 18:13, Kjetil Halvorsen wrote:>>> ?save.image >> >> And at this point it has been running with one cpu at 100% for over an >> hour! > > > It's OK to take an hour (due to memory <-> disc IO) if it uses swap space > heavily. Factor of 60 is not much given memory is faster than harddiscs by > orders of magnitude. > > UweIt takes much more than an hour! I started anew a process with the problem yesterday aroun 18.00, had to kill it this morning around 09.00. That's more than 1|5 hours. Kjetil>------------------------------ Message: 26 Date: Mon, 22 Nov 2010 10:13:01 -0500 From: Mike Marchywka <marchywka at hotmail.com> To: <kjetilbrinchmannhalvorsen at gmail.com>, <ligges at statistik.tu-dortmund.de> Cc: r-help at r-project.org Subject: Re: [R] ?summaryRprof running at 100% cpu for one hour ... Message-ID: <BLU113-W11A7426FDE41EECEB92F7BBE3D0 at phx.gbl> Content-Type: text/plain; charset="iso-8859-1" ----------------------------------------> Date: Mon, 22 Nov 2010 12:03:54 -0300 > From: kjetilbrinchmannhalvorsen at gmail.com > To: ligges at statistik.tu-dortmund.de > CC: r-help at r-project.org > Subject: Re: [R] ?summaryRprof running at 100% cpu for one hour ... > > see below. > > 2010/11/21 Uwe Ligges : > > > > > > On 21.11.2010 18:13, Kjetil Halvorsen wrote: > > >>> ?save.image > >> > >> And at this point it has been running with one cpu at 100% for over an > >> hour! > > > > > > It's OK to take an hour (due to memory <-> disc IO) if it uses swapspace> > heavily. Factor of 60 is not much given memory is faster than harddiscsby> > orders of magnitude. > > > > Uwe > > It takes much more than an hour! I started anew a process with the > problem yesterday aroun 18.00, had to kill it this morning around > 09.00. That's more than 1|5 hours.Again, see if you can run it under gdb or at least look at tools you have to determine page faults. My brain has been corrupted with 'dohs but in task manager CPU usage drops when page faults start or lock startvation etc. A blocking thread should yield IIRC. Waiting for it to die a natural death may not be practical. I just posted something on this after following another's suggestion but it should be easy for you to get developer tools, execute gdb, point it at R and then break a few times. Debuggers don't speed anything up but presumably it gets into its limit cycle ( infinite futile loop ) within a short time. Also sometimes you get these loops due to memory corruption with native code etc etc so confusing results may take a few different approaches to figure out. Turning on profiling will at best destry any memory coherence and worse ad to VM thrashing. At least try to determine if you are faulting all over.> > Kjetil > > > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 27 Date: Mon, 22 Nov 2010 09:13:19 -0600 From: Mike Lawrence <Mike.Lawrence at dal.ca> To: "r-help at lists.R-project.org" <r-help at r-project.org> Subject: [R] hierarchical mixture modelling Message-ID: <AANLkTi=fzD+Jz1-DnOY239U3xycCRfriyZ1bF-AJjjAh at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Hi folks, I have circular data that I'd like to model as a mixture of a uniform and a Von Mises centered on zero. The data are from 2 groups of human participants, where each participant is measured repeatedly within each of several conditions. So, the data frame would look something like: ######## design = expand.grid( person = 1:20 , variable1 = 1:2 , variable2 = 1:3 , repetition = 1:100 ) design$group = design$person %% 2 ######## where each row would have a data point. Now, I know how to fit the mixture of a uniform and Von Mises to each individual cell of the data independently using the EM algorithm, yielding estimates of the mixture proportion and Von Mises concentration per cell. However, I of course want to assess the degree to which the group and other variables in the design affect these model parameters, and at least in the case of the proportion estimate, I'm uncomfortable submitting the raw proportion to a test that is going to assume Gaussian error (eg. ANOVA, or lmer(..., family=gaussian)). I'm aware that lmer lets one specify non-gaussian links, but as I understand it, if I wanted to, say, specify the binomial link (which seems appropriate for a proportion), lmer wants the data to be the raw 1's and 0's, not the proportion estimate obtained from EM. I've heard that there are hierarchical mixture modelling methods out there (possibly Bayesian hierarchical mixture modelling) that might let me achieve model fitting and inference in one step (eg. model the mixture and influence on each parameter from the between and within-person variables, and treating people as random effects), but I'm having trouble tacking down instructions on how to do this. [[elided Yahoo spam]] Cheers, Mike -- Mike Lawrence Graduate Student Department of Psychology Dalhousie University Looking to arrange a meeting? Check my public calendar: http://tr.im/mikes_public_calendar ~ Certainty is folly... I think. ~ ------------------------------ Message: 28 Date: Mon, 22 Nov 2010 20:44:51 +0530 From: "Santosh Srinivas" <santosh.srinivas at gmail.com> To: <r-help at r-project.org> Subject: [R] Check for is.object Message-ID: <4cea88f0.1f44960a.1cfe.fffff56a at mx.google.com> Content-Type: text/plain; charset="us-ascii" Hello, I am trying to recursively append some data from multiple files into a common object For this, I am using in a loop NewObject <- rbind(NewObject,tempObject) For the first loop, obviously there is no NewObject ... so I wanted to do NewObject <- tempObject[0,] Now when it loops again I want to put the statement do "NewObject <- tempObject[0,]" inside a if statement ... so that it does I can skip it once NewObject has been initialized. But, is.object doesn't seem to work. What is the alternative check that I can do? And is there a better way to achieve what I want? Thanks, S ------------------------------ Message: 29 Date: Mon, 22 Nov 2010 07:17:22 -0800 (PST) From: Wonsang You <you at ifn-magdeburg.de> To: r-help at r-project.org Subject: [R] Fast Two-Dimensional Optimization Message-ID: <AANLkTikBQPjKypX7Y9X7sn3h41nO=r8WOcRtbr2dEO40 at mail.gmail.com> Content-Type: text/plain Dear R Helpers, I have attempted "optim" function to solve a two-dimensional optimization problem. It took around 25 second to complete the procedure. However, I want to reduce the computation time: less than 7 second. Is there any optimization function in R which is very rapid? Best Regards, Wonsang ----- Wonsang You Leibniz Institute for Neurobiology -- View this message in context: http://r.789695.n4.nabble.com/R-Fast-Two-Dimensional-Optimization-tp3053782p 3053782.html Sent from the R help mailing list archive at Nabble.com. [[alternative HTML version deleted]] ------------------------------ Message: 30 Date: Mon, 22 Nov 2010 16:19:14 +0100 From: ?ukasz R?c?awowicz <lukasz.reclawowicz at gmail.com> To: Marcin Gomulka <mrgomel at gmail.com> Cc: r-help at r-project.org Subject: Re: [R] plotting a timeline Message-ID: <AANLkTimfz8Eo1GZouw5hf_ROvSVDzk8mgJGjLMwCb5d4 at mail.gmail.com> Content-Type: text/plain 2010/11/20 Marcin Gomulka <mrgomel at gmail.com>> I'd rather do this with a dedicated > package function ( like axis() ).Probably you have to write your own function, or tune up manually plot. plot(the_data$eventtime, abs(the_data$impact), type="h", frame.plot=FALSE, axes FALSE, xlab="",ylab="", col="grey",lwd=2,ylim=c(-2,2),xlim=c(1913,2005)) text(the_data$eventtime,the_data$impact+.1, the_data$label,cex=.6,adj=1) lines(x=c(1914,2003),y=c(0,0),lwd=2,col="blue",t="l") axis(1,the_data$eventtime,pos=0,cex.axis=.5,padj=-2,tck=-.01) -- Mi3ego dnia [[alternative HTML version deleted]] ------------------------------ Message: 31 Date: Mon, 22 Nov 2010 07:24:56 -0800 From: Patrick Leyshock <ngkbr8es at gmail.com> To: Uwe Ligges <ligges at statistik.tu-dortmund.de>, r-help at r-project.org Subject: Re: [R] memory profiling Message-ID: <AANLkTim62ptR2idZzbLDTZ5Mtm31_09pzSAyjT6EkK_1 at mail.gmail.com> Content-Type: text/plain> Using: > > summaryRprof(memory="both") > > did the trick, thank you. I had not been using that setting when calling > summaryRprof. > > Thanks, Patrick > > 2010/11/20 Uwe Ligges <ligges at statistik.tu-dortmund.de> > > >> >> On 19.11.2010 21:48, Patrick Leyshock wrote: >> >>> I'm trying to configure Version 2.12.0 or R to do memory profiling. >>> >>> I've reconfigured the code: >>> >>> % ./compile --enable-memory-profiling=YES >>> >>> and verified that it's configured correctly by examining the output. I >>> then >>> rebuild R: >>> >>> % make >>> >>> Then I fire up R and run a script, using Rprof with the memory-profiling >>> switch set to TRUE: >>> >>> Rprof("output", memory.profiling=TRUE); >>> # a bunch of R code >>> Rprof(NULL); >>> >> >> >> Wen I do >> >> summaryRprof(memory="both") >> >> I see an additional column ... >> >> but since you have not said what you tried exactly, we cannot help very >> much. >> >> Uwe Ligges >> >> >> >> When I examine the output, however, using either R CMD Rprof from the >>> shell, >>> or summaryRprof from within R, the output I see is identical to the >>> output I >>> got when I ran R BEFORE I recompiled with memory profiling enabled. >>> >>> Anyone see something that I'm missing? >>> >>> Thanks, Patrick >>> >>> [[alternative HTML version deleted]] >>> >>> ______________________________________________ >>> R-help at r-project.org mailing list >>> https://stat.ethz.ch/mailman/listinfo/r-help >>> PLEASE do read the posting guide >>> http://www.R-project.org/posting-guide.html >>> and provide commented, minimal, self-contained, reproducible code. >>> >> >[[alternative HTML version deleted]] ------------------------------ Message: 32 Date: Mon, 22 Nov 2010 12:41:06 -0300 From: Kjetil Halvorsen <kjetilbrinchmannhalvorsen at gmail.com> To: Mike Marchywka <marchywka at hotmail.com> Cc: r-help at r-project.org, ligges at statistik.tu-dortmund.de Subject: Re: [R] ?summaryRprof running at 100% cpu for one hour ... Message-ID: <AANLkTimevZO2TA=5tZj92eZBwd5pc7GaNzRXfO=Rskx=@mail.gmail.com> Content-Type: text/plain; charset=UTF-8 see below. On Mon, Nov 22, 2010 at 12:13 PM, Mike Marchywka <marchywka at hotmail.com> wrote:> > > > > > ---------------------------------------- >> Date: Mon, 22 Nov 2010 12:03:54 -0300 >> From: kjetilbrinchmannhalvorsen at gmail.com >> To: ligges at statistik.tu-dortmund.de >> CC: r-help at r-project.org >> Subject: Re: [R] ?summaryRprof running at 100% cpu for one hour ... >> >> see below. >> >> 2010/11/21 Uwe Ligges : >> > >> > >> > On 21.11.2010 18:13, Kjetil Halvorsen wrote: >> >> >>> ?save.image >> >> >> >> And at this point it has been running with one cpu at 100% for over an >> >> hour! >> > >> > >> > It's OK to take an hour (due to memory <-> disc IO) if it uses swapspace>> > heavily. Factor of 60 is not much given memory is faster than harddiscsby>> > orders of magnitude. >> > >> > Uwe >> >> It takes much more than an hour! I started anew a process with the >> problem yesterday aroun 18.00, had to kill it this morning around >> 09.00. That's more than 1|5 hours. > > > Again, see if you can run it under gdb or at least look at > tools you have to determine page faults. My brain has been corrupted > with 'dohs but in task manager CPU usage drops when page faults start > or lock startvation etc. A blocking thread should yield IIRC. Waiting > for it to die a natural death may not be practical. >Thanks. Will try. Really, I tried yesterday, to run R under gdb within emacs, but it did'nt work out. What I did (in emacs 23) was, typing Ctrl-u M-x R and then enter the option --debugger=gdb [[elided Yahoo spam]] Kjetil> I just posted something on this after following another's suggestion but > it should be easy for you to get developer tools, execute gdb, > point it at R and then break a few times. Debuggers don't speed anything > up but presumably it gets into its limit cycle ( infinite futile loop ) > within a short time. Also sometimes you get these loops due to memorycorruption> with native code etc etc so confusing results may take a few differentapproaches> to figure out. > > Turning on profiling will at best destry any memory coherence and worse > ad to VM thrashing. At least try to determine if you are faulting allover.> > >> >> Kjetil >> > >> >> ______________________________________________ >> R-help at r-project.org mailing list >> https://stat.ethz.ch/mailman/listinfo/r-help >> PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html>> and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 33 Date: Mon, 22 Nov 2010 16:44:27 +0100 From: Martin Maechler <maechler at stat.math.ethz.ch> Cc: r-help at r-project.org Subject: Re: [R] An empty grey diagram Message-ID: <19690.36827.343657.468114 at lynne.math.ethz.ch> Content-Type: text/plain; charset=us-ascii>>>>> on Sat, 20 Nov 2010 00:12:07 -0800 (PST) writes:> Hi Josh and David, > Problem solved. > Both following steps work. > 1) >> ToothGrowth >> attach(ToothGrowth) >> plot(dose,len) # this step is needed. Don't close the diagram. Otherwise >> following command won't work. >> matrics=lm(len~dose) >> abline(matrics) > Graph displayed > 2) >> ToothGrowth >> attach(ToothGrowth) >> plot(dose, len) # this step is needed. Don't close the diagram. Otherwise >> following command won't work. >> abline(lm(len ~ dose, data = ToothGrowth)) > Graph displayed Well, it is *VERY BAD* style and error-prone practice nowadays, to attach a data set. [The only thing to attach() are save()d data *files*; there it's often tidier to attach() instead of to load() ...] I have not followed all the things you tried .. or did not try. A much better way of achieving the above (and yes; *never* close the graphics window within these) should be plot ( len ~ dose, data=ToothGrowth) abline(lm(len ~ dose, data=ToothGrowth)) and you can see {if you use a fixed-width font in your e-mail, it "springs into your eyes"} how nicely the formula notation of graphics alings with the same in models. If the above two simple lines do not work correctly, then your R environment is "broken" in some way, and maybe first doing rm(list = ls()) may help. Martin Maechler, ETH Zurich and R Core Team > B.R. > Stephen L > ----- Original Message ---- > From: Joshua Wiley <jwiley.psych at gmail.com> > Cc: r-help at r-project.org > Sent: Sat, November 20, 2010 1:39:45 PM > Subject: Re: [R] An empty grey diagram wrote: >> Hi David, >> >> >>> What happens when you follow the directions... i.e. type: >>> plot.new() #??? >> >> abline(lm(len ~ dose, data = ToothGrowth)) >> plot.new() >> >> The grey background changes to white, still an empty graph > You cannot just use abline() on an empty graphic (well, you can but > you get an empty graph). Please actually run my code, it will create > a scatter plot, then add a line. Do not close the graphic device in > between. > with(ToothGrowth, plot(dose, len)) > abline(lm(len ~ dose, data = ToothGrowth)) > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. ------------------------------ Message: 34 Date: Mon, 22 Nov 2010 10:57:06 -0500 From: Mike Marchywka <marchywka at hotmail.com> To: <kjetilbrinchmannhalvorsen at gmail.com> Cc: r-help at r-project.org, ligges at statistik.tu-dortmund.de Subject: Re: [R] ?summaryRprof running at 100% cpu for one hour ... Message-ID: <BLU113-W27D876C2B7D93140F1135BE3D0 at phx.gbl> Content-Type: text/plain; charset="iso-8859-1" ----------------------------------------> Date: Mon, 22 Nov 2010 12:41:06 -0300 > Subject: Re: [R] ?summaryRprof running at 100% cpu for one hour ... > From: kjetilbrinchmannhalvorsen at gmail.com > To: marchywka at hotmail.com > CC: ligges at statistik.tu-dortmund.de; r-help at r-project.org > > see below. > > On Mon, Nov 22, 2010 at 12:13 PM, Mike Marchywka wrote: > > > > > Thanks. Will try. Really, I tried yesterday, to run R under gdb within > emacs, but it did'nt work out. What I did (in emacs 23) was, typing > Ctrl-u M-x R > and then enter the option > --debugger=gdb >[[elided Hotmail spam]]> > KjetilI rarely use gdb but it did seem to work with R but I executed gdb from cygwin windoh and IIRC ctrl-C worked fine as it broke into debugger. I guess you could try that- start gdb and attach or invoke R from gdb. ------------------------------ Message: 35 Date: Mon, 22 Nov 2010 11:02:55 -0500 From: Jonathan P Daily <jdaily at usgs.gov> To: "Santosh Srinivas" <santosh.srinivas at gmail.com> Cc: r-help at r-project.org, r-help-bounces at r-project.org Subject: Re: [R] Check for is.object Message-ID: <OF1BD5F51B.26988C10-ON852577E3.00581042-852577E3.005830E8 at usgs.gov> Content-Type: text/plain; charset="US-ASCII" I think you want the function ?exists>if(!exists("NewObject"))-------------------------------------- Jonathan P. Daily Technician - USGS Leetown Science Center 11649 Leetown Road Kearneysville WV, 25430 (304) 724-4480 "Is the room still a room when its empty? Does the room, the thing itself have purpose? Or do we, what's the word... imbue it." - Jubal Early, Firefly r-help-bounces at r-project.org wrote on 11/22/2010 10:14:51 AM:> [image removed] > > [R] Check for is.object > > Santosh Srinivas > > to: > > r-help > > 11/22/2010 10:17 AM > > Sent by: > > r-help-bounces at r-project.org > > Hello, > > I am trying to recursively append some data from multiple files into a > common object > > For this, I am using in a loop > > NewObject <- rbind(NewObject,tempObject) > > > For the first loop, obviously there is no NewObject ... so I wanted todo> NewObject <- tempObject[0,] > > Now when it loops again I want to put the statement do "NewObject <- > tempObject[0,]" inside a if statement ... so that it does I can skip itonce> NewObject has been initialized. > > But, is.object doesn't seem to work. > > What is the alternative check that I can do? And is there a better wayto> achieve what I want? > > Thanks, > S > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 36 Date: Mon, 22 Nov 2010 08:07:24 -0800 (PST) From: Berend Hasselman <bhh at xs4all.nl> To: r-help at r-project.org Subject: Re: [R] solve nonlinear equation using BBsolve Message-ID: <1290442044850-3053902.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii After my reply you sent me the following privately:> Thank you for your respond. Now I adjust the parameters which are close > to the value that I'm expected and it gives me the following message: I > tried a few combination. My question is why it said "Unsuccessful > convergence" but still give some answers? Can I use the answers? > > >> p0 <- c(2.5,25,25,0.8) >> mgf_gammasum(p0) > [1] -17.3600000 -3410.6900000 0.3508769 -1.1028287 >> BBsolve(par = p0, fn = mgf_gammasum)[[1]] > Unsuccessful convergence. > [1] 2.066909 44.068739 24.809270 0.972542 >> source(.trPaths[5], echo=TRUE, max.deparse.length=10000) >> p0 <- c(1.7,36,50,0.9) >> mgf_gammasum(p0) > [1] 3.8400000 2601.0300000 0.7232021 0.2866732 >> BBsolve(par = p0, fn = mgf_gammasum)[[1]] > Unsuccessful convergence. > [1] 2.0823407 18.3757502 49.9935914 0.9456666 >> p0 <- c(2,40,40,0.8) >> mgf_gammasum(p0) > [1] 17.6400000 2798.7100000 0.4883676 -0.5653881 >> BBsolve(par = p0, fn = mgf_gammasum)[[1]] > Unsuccessful convergence. > [1] 2.059853 29.215478 39.882727 0.914894 >It is only giving you the values it stopped at. You are only printing [[1]] of BBsolve's result. BBsolve provides more information. You can easily check if the result is usable. Do something like this. p0 <- c(2.5,25,25,0.8) bb.result <- BBsolve(par = p0, fn = mgf_gammasum) bb.result mgf_gammasum(bb.result$par) You will see that BBsolve has NOT found a solution. If you use nleqslv as follows you will see that the jacobian matrix in your starting point is very ill-conditioned. nleqslv(p0,mgf_gammasum) All your other starting points have similar problems. You really need to rethink your system of equations. In future please also reply to the list and not only to me privately. best Berend -- View this message in context: http://r.789695.n4.nabble.com/solve-nonlinear-equation-using-BBsolve-tp30521 67p3053902.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 37 Date: Mon, 22 Nov 2010 11:07:43 -0500 From: Ista Zahn <izahn at psych.rochester.edu> Cc: R Project Help <R-help at r-project.org> Subject: Re: [R] Alternatives to image(...) and filled.contour(...) for 2-D filled Plots Message-ID: <AANLkTi=9Z+ozud2YRj4C2Jqu5xRJGph4vSusQw4UbRHp at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Hi Jason, You do not say what you want the alternative to do, so its hard to know if this will be helpful. But one alternative is dat <- as.data.frame(ak.fan) dat <- melt(dat, id.vars=c("x", "y")) p <- ggplot(dat, aes(x=x, y=variable)) p + geom_tile(aes(fill=value)) -Ista wrote:> > By any chance are there any alternatives to image(...) andfilled.contour(...)> > I used Rseek to search for that very topic, but didn't turn over anyleads...>http://www.rseek.org/?cx=010923144343702598753%3Aboaz1reyxd4&newwindow=1&q=a lternative+to+image+and+filled.contour&sa=Search&cof=FORID%3A11&siteurl=www. rseek.org%252F#1238> > > I'm sure there are some out there, but curious about some of the favoritesand> ones folks have had success using. > > > Thanks for any insights and feedback. > > I would like to use the alternative 2-D fill function with the example Ihave> been messing with in place of image(...) or filled.contour(...): > > > > library(akima) > > hyp_distance<-seq(1,15) > angle_deg_val<-seq(0,15) > > > x_distance_val<-NULL > y_distance_val<-NULL > > for(ii in 1:length(hyp_distance)) > { > ? ? ? ?for(jj in 1:length(angle_deg_val)) > ? ? ? ?{ > ? ? ? ? ? ? ??x_distance_tmp<-hyp_distance[ii]*cos(angle_deg_val[jj]*pi/180)> ? ? ? ? ? ? ??y_distance_tmp<-hyp_distance[ii]*sin(angle_deg_val[jj]*pi/180)> > ? ? ? ? ? ? ? ?x_distance_val<-c(x_distance_val, x_distance_tmp) > ? ? ? ? ? ? ? ?y_distance_val<-c(y_distance_val, y_distance_tmp) > ? ? ? ?} > > } > > > temperature_vals<-rnorm(length(x_distance_val), 75, 2) > > temp_samples<-cbind(x_distance_val, y_distance_val, temperature_vals) > > temp_samples_DF<-data.frame(x = x_distance_val, y = ?y_distance_val, z > temperature_vals) > > > ak.fan <- interp(temp_samples[,1], temp_samples[,2], temp_samples[,3] ) > > length_val<-floor(max(temperature_vals) - min(temperature_vals))*2 > > color_vals_red_to_yellow_to_green<-colorRampPalette(c("red", "yellow","green"),> space="Lab")(length_val) > color_vals_green_to_yellow_to_red<-colorRampPalette(c("green", "yellow","red"),> space="Lab")(length_val) > > plot(1,1, col = 0, xlim = c(min(x_distance_val), max(x_distance_val)),ylim > c(min(y_distance_val), max(y_distance_val)), xlab = "Room X Position (FT)", ylab> = "Room Y Position (FT)", main = "Room Temp vs Position") > > grid() > > # filled.contour(ak.fan, col = color_vals_red_to_yellow_to_green) > # filled.contour(ak.fan, col = color_vals_green_to_yellow_to_red) > > # image(ak.fan, col = color_vals_red_to_yellow_to_green, add = TRUE) > image(ak.fan, col = color_vals_green_to_yellow_to_red, add = TRUE) > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >-- Ista Zahn Graduate student University of Rochester Department of Clinical and Social Psychology http://yourpsyche.org ------------------------------ Message: 38 Date: Mon, 22 Nov 2010 11:09:10 -0500 From: Matt Shotwell <shotwelm at musc.edu> To: "martin.tomko at geo.uzh.ch" <martin.tomko at geo.uzh.ch> Cc: "r-help at r-project.org" <r-help at r-project.org> Subject: Re: [R] statistical test for comparison of two classifications (nominal) Message-ID: <1290442150.1756.74.camel at matt-laptop> Content-Type: text/plain; charset="UTF-8" Martin, Pardon the delayed reply. Bootstrap methods have been around for some time (late seventies?), but their popularity seems to have exploded in correspondence with computing technology. You should be able to find more information in most modern books on statistical inference, but here is a brief: The bootstrap is a method often used to establish an empirical null distribution for a test statistic when traditional (analytical) methods fail. The bootstrap works by imposing a null hypothesis on the observed data, followed by re-sampling with replacement. The test statistic is computed at each re-sample and used to build up an empirical null distribution. The idea is to impose the null hypothesis while preserving variability in the observed data, and thus the test statistic. For example, suppose we observe some continuous scalar data and hypothesize that the sample was observed from a population with mean zero. We can impose this hypothesis by subtracting the sample mean from each observation. Re-samples from these transformed data are treated as having been observed under the null hypothesis. In the case of classification and partitioning, the difficulty is formulating a meaningful null hypothesis about the collection of classifications, and imposing the null hypothesis in a bootstrap sampling scheme. -Matt On Wed, 2010-11-17 at 10:01 -0500, Martin Tomko wrote:> Thanks Mat, > I have in the meantime identified the Rand index, but not the others. I > will also have a look at profdpm, that did not pop-up in my searches. > Indeed, the interpretation is going to be critical... Could you please > elaborate on what you mean by the bootstrap process? > > Thanks a lot for your helps, > Martin > > On 11/17/2010 3:50 PM, Matt Shotwell wrote: > > There are several statistics used to compare nominal classifications, or > > _partitions_ of a data set. A partition isn't quite the same in this > > context because partitioned data are not restricted to a fixed number of > > classes. However, the statistics used to compare partitions should also > > work for these 'restricted' partitions. See the Rand index, Fowlkes and > > Mallows index, Wallace indices, and the Jaccard index. The profdpm > > package implements a function (?profdpm::pci) that computes these > > indices for two factors representing partitions of the same data. > > > > The difficult part is drawing statistical inference about these indices. > > It's difficult to formulate a null hypothesis, and even more difficult > > to determine a null distribution for a partition comparison index. A > > bootstrap test might work, but you will probably have to implement this > > yourself. > > > > -Matt > > > > On Wed, 2010-11-17 at 08:33 -0500, Martin Tomko wrote: > > > >> Dear all, > >> I am having a hard time to figure out a suitable test for the match > >> between two nominal classifications of the same set of data. > >> I have used hierarchical clustering with multiple methods (ward, > >> k-means,...) to classify my dat into a set number of classesa, and I > >> would like to compare the resulting automated classification with the > >> actual - objective benchmark one. > >> So in principle I have a data frame with n columns of nominal > >> classifications, and I want to do a mutual comparison and test for > >> significance in difference in classification between pairs of columns. > >> > >> I just need to identify a suitable test, but I fail. I am currently > >> exploring the possibility of using Cohen's Kappa, but I am open toother> >> suggestions. Especially the fact that kappa seems to be moslty used on > >> failible, human annotators seems to bring in limitations taht do not > >> apply to my automatic classification. > >> Any help will be appreciated, especially if also followed by a pointer > >> to an R package that implements it. > >> > >> Thanks > >> Martin > >> > >> ______________________________________________ > >> R-help at r-project.org mailing list > >> https://stat.ethz.ch/mailman/listinfo/r-help > >> PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> >> and provide commented, minimal, self-contained, reproducible code. > >> > > > >-- Matthew S. Shotwell Graduate Student Division of Biostatistics and Epidemiology Medical University of South Carolina ------------------------------ Message: 39 Date: Mon, 22 Nov 2010 08:58:30 -0800 (PST) From: Phil Spector <spector at stat.berkeley.edu> To: Santosh Srinivas <santosh.srinivas at gmail.com> Cc: r-help at r-project.org Subject: Re: [R] Check for is.object Message-ID: <alpine.DEB.2.00.1011220852200.14043 at springer.Berkeley.EDU> Content-Type: TEXT/PLAIN; charset=US-ASCII; format=flowed Santosh - The simple answer to your question is to initialize NewObject to NULL before the loop, i.e. newObject = NULL However, I should point out this is one of the most inefficient ways to program in R. A far better way is to allocate enough space for NewObject outside your loop, and "fill it in" in the loop. Here's a simple example to give you an idea of the difference in time the two methods require:> system.time({answer = matrix(NA,1000,5);+ for(i in 1:1000)answer[i,] <- sample(10,5)}) user system elapsed 0.020 0.000 0.017> system.time({answer=NULL;+ for(i in 1:1000)answer=rbind(answer,sample(10,5))}) user system elapsed 0.072 0.000 0.070 However, it gets even worse if the sample size is larger:> system.time({answer = matrix(NA,10000,5);+ for(i in 1:10000)answer[i,] <- sample(10,5)}) user system elapsed 0.184 0.000 0.184> system.time({answer=NULL;for(i in 1:10000)+ answer=rbind(answer,sample(10,5))}) user system elapsed 5.492 0.032 5.562 Even if you don't know how big your newObject matrix will become, it's still far more efficient to overallocate the matrix and then truncate it at the end. I'd strongly recommend that you avoid building your matrix [[elided Yahoo spam]] - Phil Spector Statistical Computing Facility Department of Statistics UC Berkeley spector at stat.berkeley.edu On Mon, 22 Nov 2010, Santosh Srinivas wrote:> Hello, > > I am trying to recursively append some data from multiple files into a > common object > > For this, I am using in a loop > > NewObject <- rbind(NewObject,tempObject) > > > For the first loop, obviously there is no NewObject ... so I wanted to do > NewObject <- tempObject[0,] > > Now when it loops again I want to put the statement do "NewObject <- > tempObject[0,]" inside a if statement ... so that it does I can skip itonce> NewObject has been initialized. > > But, is.object doesn't seem to work. > > What is the alternative check that I can do? And is there a better way to > achieve what I want? > > Thanks, > S > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >------------------------------ Message: 40 Date: Mon, 22 Nov 2010 12:04:59 -0500 From: Ista Zahn <izahn at psych.rochester.edu> To: madr <madrazel at interia.pl> Cc: r-help at r-project.org Subject: Re: [R] "negative alpha" or custom gradient colors of data dots in scatterplot ? Message-ID: <AANLkTikVspi-vK+qeB+QmH_qvqFi2+T2deeKhqyLxjTJ at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Hi, I suggest taking a look at the plotting functions in the ggplot2 package. For example: x <- rnorm(10000) y <- x+rnorm(10000) dat <- data.frame(x,y) library(ggplot2) p <- ggplot(dat, aes(x=x, y=y)) p + geom_point() # too much overplotting: compare to dev.new() p + geom_hex(binwidth=c(.1,.1)) Best, Ista On Sun, Nov 21, 2010 at 9:13 AM, madr <madrazel at interia.pl> wrote:> > I know that by setting alpha to for example col = rgb(0, 0, 0, 0.1) it is > possible to see how many overlapping is in the plot. But disadvantage ofit> is that single points are barely visible on the background. So I wonder if > there is possible to make setting that single points would be almostblack,> but with more and more data on the same spot it would get more and more > whiteish. Or maybe it is possible to make sole data points black but > overlapped tending to some particular color of choice ? > -- > View this message in context:http://r.789695.n4.nabble.com/negative-alpha-or-custom-gradient-colors-of-da ta-dots-in-scatterplot-tp3052394p3052394.html> Sent from the R help mailing list archive at Nabble.com. > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >-- Ista Zahn Graduate student University of Rochester Department of Clinical and Social Psychology http://yourpsyche.org ------------------------------ Message: 41 Date: Mon, 22 Nov 2010 18:16:42 +0100 From: wphantomfr <wphantomfr at gmail.com> To: r-help at r-project.org Subject: [R] Problem setting the number of digits in xtable Message-ID: <8E449FB3-99DF-4BC3-A1DA-EBDC075F6D0A at gmail.com> Content-Type: text/plain DEar list members, I am currently using Sweave with LaTeX which is great. I can use xtable for formatting outp of tables but I have a problem setting the number of decimals in xtables when used with dataframe. I have found an example on the net ith matrix and it works. For example this works :> > tmp <- matrix(rnorm(9), 3, 3) > > xtmp <- xtable(tmp) > > digits(xtmp) <- c(0,0,3,4) > > print(xtmp, include.rownames = FALSE) # row namesproduced :> % latex table generated in R 2.12.0 by xtable 1.5-6 package > % Mon Nov 22 17:35:00 2010 > \begin{table}[ht] > \begin{center} > \begin{tabular}{rrr} > \hline > 1 & 2 & 3 \\ > \hline > -2 & -2.158 & 2.8886 \\ > 1 & 1.330 & 0.4677 \\ > -0 & 0.486 & -0.3319 \\ > \hline > \end{tabular} > \end{center} > \end{table}But this won't work :> > mydata > TEST t ddl p CONDITION > 2 R1 3.01109061083632 16 0.00828552765650315 C1 > 3 R2 3.30476953908811 16 0.00447412002109504 C1 > 4 DR 2.86343993410509 16 0.0112631908739966 C1 > 5 R1 1.05386387510206 16 0.30760068470456 C2 > 6 R2 3.04997140665209 16 0.00763921045771104 C2 > 7 DR 2.25175987512241 16 0.0387401575011488 C2but> > xtable(mydata,digits=2)produced> % latex table generated in R 2.12.0 by xtable 1.5-6 package > % Mon Nov 22 18:13:47 2010 > \begin{table}[ht] > \begin{center} > \begin{tabular}{rlllll} > \hline > & TEST & t & ddl & p & CONDITION \\ > \hline > 2 & R1 & 3.01109061083632 & 16 & 0.00828552765650315 & C1 \\ > 3 & R2 & 3.30476953908811 & 16 & 0.00447412002109504 & C1 \\ > 4 & DR & 2.86343993410509 & 16 & 0.0112631908739966 & C1 \\ > 5 & R1 & 1.05386387510206 & 16 & 0.30760068470456 & C2 \\ > 6 & R2 & 3.04997140665209 & 16 & 0.00763921045771104 & C2 \\ > 7 & DR & 2.25175987512241 & 16 & 0.0387401575011488 & C2 \\ > \hline > \end{tabular} > \end{center} > \end{table}I have also tried setting the digits with c(0,0,4,0,4,0), using also the 'display' argument to specify the type of each column... noway... What am I missing ? Thanks in advance Sylvain Climent [[alternative HTML version deleted]] ------------------------------ Message: 42 Date: Mon, 22 Nov 2010 18:33:07 +0100 (CET) From: omerle <omerle at laposte.net> To: r-help at r-project.org Subject: [R] RCurl : All connection are used ? Message-ID: <1277930.342.1290447187435.JavaMail.www at wwinf8202> Content-Type: text/plain B B B B B Hi everybody, I got a problem with the ftpUpload function from the RCurl package. My goal is to Upload a lot of files from a local directory to a web server. 1st try : for (i in 1:length(file)){ B B B B B ftpUpload(what=files[i],to=files[i]) } At i=11 I get : (my server has only 10 available open connections available) : Erreur dans curlPerform(url = to, upload = TRUE, readfunction uploadFunctionHandler(what,B : B Got a 421 ftp-server response when 220 was expected 2 nd Try : ftpConnection=getCurlHandle(userpwd=ftp$userpwd,maxconnects=1,fresh.connect0) for (i in 1:length(file)){ B B B B B ftpUpload(what=files[i],to=files[i],curl=ftpConnection) } And I got this error after 30 files (the error is not linked to the web server but to the R session) : Error in file(file, "rb") : all conection are used I read the documentation and the options from curl library but I can't find how to solve my problem even if I think that the problem is linked to not closing the position I opened. Do you have any idea how to solve it ? Thanks, Olivier Merle Une messagerie gratuite, garantie C vie et des services en plus, C'a vous tente ? Je crC)e ma boC.te mail www.laposte.net [[alternative HTML version deleted]] ------------------------------ Message: 43 Date: Mon, 22 Nov 2010 18:47:16 +0100 From: Lucia Ca?as <lucia.canas at co.ieo.es> To: <r-help at R-project.org> Subject: [R] sm.ancova graphic Message-ID: <50EB6473669C6741AC34FF5692DCC4560DA79E at ieocoruna.co.ieo.es> Content-Type: text/plain Hi R-Users, I am working with sm.ancova (in the package sm) and I have two problems with the graph, which is automatically generated when sm.ancova() is run. 1-Besides of the fitted lines, the observed data appeared automatically in the graph. I prefer that only fitted lines appear. I check the sm.options, but I could not find the way that the observed data do not appear in the graph. 2-I would like to change the size of the numbers in the axis. Again, I check the sm.options, but I could not find the correct way. Thank you in advance, Lucma [[alternative HTML version deleted]] ------------------------------ Message: 44 Date: Mon, 22 Nov 2010 18:51:04 +0100 From: JiHO <jo.lists at gmail.com> To: R Help <r-help at stat.math.ethz.ch> Subject: [R] Plotting a cloud/fog of variable density in rgl Message-ID: <AANLkTi=eYZV6xjQdKHnP1ak0yO4VQDrPMp_hRpMX=d6Z at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 Hi everyone, I want to plot a 3D interpolation of the concentration of aquatic organisms. My goal would be to have the result represented as clouds with a density proportional to the abundance of organisms, so that I could fly (well, swim actually ;) ) through the scene and see the patches here and there. Basically, I want to do something like this: http://www.youtube.com/watch?v=27mo_Y-aU-c but simpler and with only clouds. I though about doing it this way: 1- interpolate to a fine grid 2- plot points at each grid intersection of transparency inversely proportional to abundance 3- blur/fog a bit each point to create the general impression of a cloud So far I am stuck on 3 but maybe there is a better overall solution. Here is some code that reads the result of the interpolation on a coarse grid and plots it: # read a set of gridded data points in 3D d = read.table("http://dl.dropbox.com/u/1047321/R/test3Ddata.txt", header=T) # plot library("rgl") spheres3d(d$x, d$y, d$z, alpha=alpha, radius=0.05) And here is a version that actually performs the interpolation a random set of points in 3D through kriging in case you want to try with increase precision. # create a set of random data points in 3D n = 50 data3D = data.frame(x = runif(n), y = runif(n), z = runif(n), v rnorm(n)) # do 3d interpolation via kriging library("gstat") coordinates(data3D) = ~x+y+z range1D = seq(from = 0, to = 1, length = 10) grid3D = expand.grid(x = range1D, y = range1D, z = range1D) gridded(grid3D) = ~x+y+z res3D = krige(formula = v ~ 1, data3D, grid3D, model = vgm(1, "Exp", .2)) # convert the result to a data.frame d = as.data.frame(res3D) # compute transparency (proportional to the interpolated value) maxD = max(d$var1.pred) minD = min(d$var1.pred) alpha = (d$var1.pred - minD)/(maxD - minD) # reduce maximum alpha (all points are semi-transparent) alpha = alpha/5 # plot library("rgl") spheres3d(d$x, d$y, d$z, alpha=alpha, radius=0.05) I saw the fog effect but it seems to add a fog in the scene to increase depth. What I want is my scene to actually look like a fog. Thanks in advance for any help. Sincerely, JiHO --- http://maururu.net ------------------------------ Message: 45 Date: Mon, 22 Nov 2010 13:13:55 -0500 From: Mike Marchywka <marchywka at hotmail.com> To: <omerle at laposte.net>, <r-help at r-project.org> Subject: Re: [R] RCurl : All connection are used ? Message-ID: <BLU113-W1133131680D4E669D5227EBE3D0 at phx.gbl> Content-Type: text/plain; charset="iso-8859-1" I guess I would just comment that for many tasks I try to keep the work in dedicated tools. In this case, command line versions of curl or even wget. The reason is things like this that come up talking to foreign entitites. Also the learning curve can be amortized over many other efforts that require same file transfers etc. I just mention this for rebuttals from R experts. ---------------------------------------- From: omerle at laposte.net To: r-help at r-project.org Date: Mon, 22 Nov 2010 18:33:07 +0100 Subject: [R] RCurl : All connection are used ? Hi everybody, I got a problem with the ftpUpload function from the RCurl package. My goal is to Upload a lot of files from a local directory to a web server. 1st try : for (i in 1:length(file)){ ftpUpload(what=files[i],to=files[i]) } At i=11 I get : (my server has only 10 available open connections available) : Erreur dans curlPerform(url = to, upload = TRUE, readfunction uploadFunctionHandler(what, : Got a 421 ftp-server response when 220 was expected 2 nd Try : ftpConnection=getCurlHandle(userpwd=ftp$userpwd,maxconnects=1,fresh.connect0) for (i in 1:length(file)){ ftpUpload(what=files[i],to=files[i],curl=ftpConnection) } And I got this error after 30 files (the error is not linked to the web server but to the R session) : Error in file(file, "rb") : all conection are used I read the documentation and the options from curl library but I can't find how to solve my problem even if I think that the problem is linked to not closing the position I opened. Do you have any idea how to solve it ? Thanks, Olivier Merle Une messagerie gratuite, garantie ? vie et des services en plus, ?a vous tente ? Je cr?e ma bo?te mail www.laposte.net [[alternative HTML version deleted]] ______________________________________________ R-help at r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. ------------------------------ Message: 46 Date: Mon, 22 Nov 2010 10:35:02 -0800 (PST) From: Manta <mantino84 at libero.it> To: r-help at r-project.org Subject: Re: [R] Ordeing Zoo object Message-ID: <1290450902168-3054192.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii And how about if I want to order the series from the smallest to the largest value, keeping the date index in order to see when the values were predominantly negative etc... Thanks, Marco -- View this message in context: http://r.789695.n4.nabble.com/Ordeing-Zoo-object-tp955868p3054192.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 47 Date: Mon, 22 Nov 2010 13:55:13 -0500 From: Duncan Murdoch <murdoch.duncan at gmail.com> To: JiHO <jo.lists at gmail.com> Cc: R Help <r-help at stat.math.ethz.ch> Subject: Re: [R] Plotting a cloud/fog of variable density in rgl Message-ID: <4CEABC91.1080402 at gmail.com> Content-Type: text/plain; charset=ISO-8859-1; format=flowed On 22/11/2010 12:51 PM, JiHO wrote:> Hi everyone, > > I want to plot a 3D interpolation of the concentration of aquatic > organisms. My goal would be to have the result represented as clouds > with a density proportional to the abundance of organisms, so that I > could fly (well, swim actually ;) ) through the scene and see the > patches here and there. Basically, I want to do something like this: > http://www.youtube.com/watch?v=27mo_Y-aU-c > but simpler and with only clouds.rgl doesn't make everything in OpenGL available. I'm not sure exactly how those clouds were done, but it wouldn't really be easy to do them in rgl. I think you can come closest to what you want within rgl by using sprites rather than rendering transparent spheres. See examples(sprites3d). Duncan Murdoch> I though about doing it this way: > 1- interpolate to a fine grid > 2- plot points at each grid intersection of transparency inversely > proportional to abundance > 3- blur/fog a bit each point to create the general impression of a cloud > > So far I am stuck on 3 but maybe there is a better overall solution. > Here is some code that reads the result of the interpolation on a > coarse grid and plots it: > > # read a set of gridded data points in 3D > d = read.table("http://dl.dropbox.com/u/1047321/R/test3Ddata.txt",header=T)> > # plot > library("rgl") > spheres3d(d$x, d$y, d$z, alpha=alpha, radius=0.05) > > And here is a version that actually performs the interpolation a > random set of points in 3D through kriging in case you want to try > with increase precision. > > # create a set of random data points in 3D > n = 50 > data3D = data.frame(x = runif(n), y = runif(n), z = runif(n), v rnorm(n)) > > # do 3d interpolation via kriging > library("gstat") > coordinates(data3D) = ~x+y+z > range1D = seq(from = 0, to = 1, length = 10) > grid3D = expand.grid(x = range1D, y = range1D, z = range1D) > gridded(grid3D) = ~x+y+z > res3D = krige(formula = v ~ 1, data3D, grid3D, model = vgm(1, "Exp",.2))> > # convert the result to a data.frame > d = as.data.frame(res3D) > > # compute transparency (proportional to the interpolated value) > maxD = max(d$var1.pred) > minD = min(d$var1.pred) > alpha = (d$var1.pred - minD)/(maxD - minD) > # reduce maximum alpha (all points are semi-transparent) > alpha = alpha/5 > > # plot > library("rgl") > spheres3d(d$x, d$y, d$z, alpha=alpha, radius=0.05) > > > I saw the fog effect but it seems to add a fog in the scene to > increase depth. What I want is my scene to actually look like a fog. > > Thanks in advance for any help. Sincerely, > > JiHO > --- > http://maururu.net > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 48 Date: Mon, 22 Nov 2010 12:01:26 -0700 From: Greg Snow <Greg.Snow at imail.org> To: Sonja Klein <sonja.klein.07 at aberdeen.ac.uk>, "r-help at r-project.org" <r-help at r-project.org> Subject: Re: [R] How to produce a graph of glms in R? Message-ID: <B37C0A15B8FB3C468B5BC7EBC7DA14CC633FEC591B at LP-EXMBVS10.CO.IHC.COM> Content-Type: text/plain; charset="us-ascii" Look at Predict.Plot (and possibly TkPredict) in the TeachingDemos package. -- Gregory (Greg) L. Snow Ph.D. Statistical Data Center Intermountain Healthcare greg.snow at imail.org 801.408.8111> -----Original Message----- > From: r-help-bounces at r-project.org [mailto:r-help-bounces at r- > project.org] On Behalf Of Sonja Klein > Sent: Saturday, November 20, 2010 4:52 AM > To: r-help at r-project.org > Subject: [R] How to produce a graph of glms in R? > > > I'm very new to R and modeling but need some help with visualization of > glms. > > I'd like to make a graph of my glms to visualize the different effects > of > different parameters. > I've got a binary response variable (bird sightings) and use binomial > glms. > The 'main' response variable is a measure of distance to a track and > the > parameters I'm testing for are vegetation parameters that effect the > response in terms of distance. > My glm is: glm(Response~NEdist+I(NEdist^2)+Distance+I(Distance^2) which > is > the basic model and where I add interactions to, like for exampls > Visibility > as an interaction to Distance > (glm(Response~NEdist+I(NEdist^2)+Distance*Visibility+I(Distance^2))) > > I'd now like to make a graph which has the response variable on the y- > axis > (obviously). But the x-axis should have distance on it. The NEdist is a > vector that is just co-influencing the curve and has to stay in the > model > but doesn't have any interactions with any other vectors. > I'd then like to put in curves/lines for the different models to see if > for > example visibility effects the distance of the track to the first bird > sighting. > > Is there a way to produce a graph in R that has these features? > -- > View this message in context: http://r.789695.n4.nabble.com/How-to- > produce-a-graph-of-glms-in-R-tp3051471p3051471.html > Sent from the R help mailing list archive at Nabble.com. > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide http://www.R-project.org/posting- > guide.html > and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 49 Date: Mon, 22 Nov 2010 14:05:52 -0500 From: Mike Marchywka <marchywka at hotmail.com> To: <murdoch.duncan at gmail.com>, <jo.lists at gmail.com> Cc: r-help at stat.math.ethz.ch Subject: Re: [R] Plotting a cloud/fog of variable density in rgl Message-ID: <BLU113-W6EEA06C04558173792A5ABE3D0 at phx.gbl> Content-Type: text/plain; charset="iso-8859-1" ----------------------------------------> Date: Mon, 22 Nov 2010 13:55:13 -0500 > From: murdoch.duncan at gmail.com > To: jo.lists at gmail.com > CC: r-help at stat.math.ethz.ch > Subject: Re: [R] Plotting a cloud/fog of variable density in rgl > > On 22/11/2010 12:51 PM, JiHO wrote: > > Hi everyone, > > > > I want to plot a 3D interpolation of the concentration of aquatic > > organisms. My goal would be to have the result represented as clouds > > with a density proportional to the abundance of organisms, so that I > > could fly (well, swim actually ;) ) through the scene and see the > > patches here and there. Basically, I want to do something like this: > > http://www.youtube.com/watch?v=27mo_Y-aU-c > > but simpler and with only clouds. > > rgl doesn't make everything in OpenGL available. I'm not sure exactly > how those clouds were done, but it wouldn't really be easy to do them in > rgl. > > I think you can come closest to what you want within rgl by using > sprites rather than rendering transparent spheres. See > examples(sprites3d). >If you only have 2 things with simple properties, namely point emitters as your organisms and a uniform concsntration of transparent scatters ( the fog) you can probably derive geometrical optics expressions for the ray trace results and just integrate those over your source distribution. This should be reasonably easy in R. I haven't been to siggraph since 1983 so can't help much but you can probably find analyitcal solutions for fog on google and just sum up your source distribution. I guess you could even do some wave optics etc as presumably the fog could be done as a function of wavelength just as easily. In any case, if you only have two basic things with simple disto should be reasonably easy to do in R with your own code. ------------------------------ Message: 50 Date: Mon, 22 Nov 2010 15:28:29 +0100 From: "B.-Markus Schuller" <b.markus.schuller at googlemail.com> To: r-help at r-project.org Subject: Re: [R] data acquisition with R? Message-ID: <4CEA7E0D.2090606 at googlemail.com> Content-Type: text/plain; charset=ISO-8859-1; format=flowed Thanks a lot, Matt! I will have a look at the options you suggested. Cheers, Mango -- ----------------------------------------- Never run for the bus. Never skip tea. ------------------------------ Message: 51 Date: Mon, 22 Nov 2010 10:23:07 -0800 (PST) From: bogdanno <bodinsoul at gmail.com> To: r-help at r-project.org Subject: [R] Is it possible to make a matrix to start at row 0? Message-ID: <ef7292d4-2217-4646-97ce-0e8d53c732c9 at n30g2000vbb.googlegroups.com> Content-Type: text/plain; charset=ISO-8859-1 I want to make the matrix to be indexed from row (column) 0, not 1 Can I do that? How? Thanks ------------------------------ Message: 52 Date: Mon, 22 Nov 2010 16:15:17 -0200 From: csrabak <crabak at acm.org> To: r-help at stat.math.ethz.ch Subject: Re: [R] Rexcel Message-ID: <icebvm$qdh$1 at dough.gmane.org> Content-Type: text/plain; charset=ISO-8859-1; format=flowed Em 22/11/2010 10:11, Luis Felipe Parra escreveu:> Hello I am new to RExcel and I would like to run a source code form the > excel worksheet. I would like to run the following code > > source("C:\\Quantil Aplicativos\\Genercauca\\BackwardSelectionNC.r") > > from the excel wroksheet. Does anybody know how to do this? > > Thank you >Felipe, Look at the section "Startup" in the RExcel help. In a nutshell, if you want the code to run immediately at the loading of the spreadsheet, create a workbook called "RCode" and put your source there. Other options are available. See the docs. -- Cesar Rabak ------------------------------ Message: 53 Date: Mon, 22 Nov 2010 08:10:07 -0800 (PST) From: "dhacademic at gmail.com" <dhacademic at gmail.com> To: r-help at r-project.org Subject: Re: [R] question about constraint minimization Message-ID: <AANLkTi=21mXKG4dEskMMN66xt0n_GYxT7dzc_ceb1f3m at mail.gmail.com> Content-Type: text/plain Hi, I have struggled on this "bound optimization with equality constraint" by using optim function for two days, but still fail to prepare a good input. Can anyone help to prepare the input for my specific case? Many thanks. Best, Hao On Sat, Nov 20, 2010 at 3:17 AM, Hans W Borchers [via R] < ml-node+3051338-309339578-202837 at n4.nabble.com<ml-node%2B3051338-309339578-2 02837 at n4.nabble.com>> wrote:> dhacademic <at> gmail.com <dhacademic <at> gmail.com> writes: > > > > > > > Hi, > > > > I am a beginner of R. There is a question about constraint minimization. > A > > function, y=f(x1,x2,x3....x12), needs to be minimized. There are 3 > > requirements for the minimization: > > > > (1) x2+x3+...+x12=1.5 (x1 is excluded); > > (2) x1=x3=x4; > > (3) x1, x3 and x5 are in the range of -1~0, respectively. The rest > variables > > (x2, x4, x6, x7, ...., x12) are in the range of 0~1, respectively. > > > > The "optim" function is used. And part of my input is as follow, where > > "xx1r" represents the x12: > > > > xx1r=1.5-x[2]-x[1]-x[1]-x[3]-x[4]-x[5]-x[6]-x[7]-x[8]-x[9] > > start=rnorm(9) > > up=1:9/1:9*1 > > lo=1:9/1:9*-1 > > out=optim(start,f,lower=lo,upper=up,method="L-BFGS-B",hessian=TRUE, > > control=list(trace=6,maxit=1000)) > > > > There are two problems in this input. the "up" and "lo" only define a > range > > of -1~1 for x1 to x11, which can not meet the requirement (3). In > addition, > > there is not any constraint imposed on x12. I have no idea how tospecify> a > > matrix that can impose different constraints on individual variables ina> > > function. Any suggestion is highly appreciated. > > > > Best, > > Hao > > > > I don't see any direct need for real 'constraint' optimization here, > it is a 'bounded' optimization where you are allowed to use > > lower <- c(-1,0,-1,0,-1,0,0,0,0,0,0,0) > upper <- c( 0,1, 0,0, 0,1,1,1,1,1,1,1) > > Otherwise, your description is confusing: > (1) Did you change f to a new function with 9 variables, eliminating > x3, x4, and x12 ? > (2) x4 (being equal to x1) has to be in [-1, 0] but also in [0, 1]? > (3) If you need to restrict x12 to [0, 1] also, you cannot eliminate it. > Either keep x12 and use an equality constraint, or use inequality > constraints on xxlr. > > Hans Werner > > ______________________________________________ > [hidden email]<http://user/SendEmail.jtp?type=node&node=3051338&i=0>mailing list> https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. > > > ------------------------------ > View message @ >http://r.789695.n4.nabble.com/question-about-constraint-minimization-tp30508 80p3051338.html> > To unsubscribe from question about constraint minimization, clickhere<http://r.789695.n4.nabble.com/template/NamlServlet.jtp?macro=unsubscrib e_by_code&node=3050880&code=ZGhhY2FkZW1pY0BnbWFpbC5jb218MzA1MDg4MHwtNjM2Nzc0 NA==>.> >-- View this message in context: http://r.789695.n4.nabble.com/question-about-constraint-minimization-tp30508 80p3053912.html Sent from the R help mailing list archive at Nabble.com. [[alternative HTML version deleted]] ------------------------------ Message: 54 Date: Mon, 22 Nov 2010 01:41:42 -0800 (PST) From: Dimitri Shvorob <dimitri.shvorob at gmail.com> To: r-help at r-project.org Subject: Re: [R] Lost in POSIX Message-ID: <1290418902451-3053329.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii> Nor would I call this much of an improvement in clarity... what about"min"? You want to know the minimum? LOL. (And apologies for the insensitivity). Thank you for help, Jeff. This works, but I am still curious to see a solution based on "trunc", if anyone can find it. -- View this message in context: http://r.789695.n4.nabble.com/Lost-in-POSIX-tp3052768p3053329.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 55 Date: Mon, 22 Nov 2010 11:19:29 +0000 From: Georg Otto <gwo at well.ox.ac.uk> To: r-help at stat.math.ethz.ch Subject: Re: [R] Find in R and R books Message-ID: <xotr5edbo72.fsf at well.ox.ac.uk> Content-Type: text/plain; charset=us-ascii> Also when I try to search in google using for example the word R insidethe search lemma I get very few results as the R confuses the search engine. When I was looking something in matlab ofcourse it was easier to get results as the search engine performs better.> What are your tricks when you want to find some function that providessome functionality? To search R-specific sites the best place to go is this one: http://www.rseek.org/ Cheers, Georg ------------------------------ Message: 56 Date: Mon, 22 Nov 2010 00:20:09 -0800 (PST) From: meytar <meytar at techunix.technion.ac.il> To: r-help at r-project.org Subject: Re: [R] using rpart with a tree misclassification condition Message-ID: <1290414009182-3053230.post at n4.nabble.com> Content-Type: text/plain [[elided Yahoo spam]] Yes, I meant the apparent error rate. According to your advice, if I use "rpart" to build a full tree, what pruning command will be appropriate and will able me to add as an input to the pruning procedure the total error rate i'm looking for? Thank you very much Meytar -- View this message in context: http://r.789695.n4.nabble.com/using-rpart-with-a-tree-misclassification-cond ition-tp3053167p3053230.html Sent from the R help mailing list archive at Nabble.com. [[alternative HTML version deleted]] ------------------------------ Message: 57 Date: Mon, 22 Nov 2010 07:23:25 -0800 From: Patrick Leyshock <pleyshock at gmail.com> To: Uwe Ligges <ligges at statistik.tu-dortmund.de> Cc: r-help at r-project.org Subject: Re: [R] memory profiling Message-ID: <AANLkTinWYjpGPjFre3Q43rNi1x3ZgLYG7Y6wAQcqzd31 at mail.gmail.com> Content-Type: text/plain Using: summaryRprof(memory="both") did the trick, thank you. I had not been using that setting when calling summaryRprof. Thanks, Patrick 2010/11/20 Uwe Ligges <ligges at statistik.tu-dortmund.de>> > > On 19.11.2010 21:48, Patrick Leyshock wrote: > >> I'm trying to configure Version 2.12.0 or R to do memory profiling. >> >> I've reconfigured the code: >> >> % ./compile --enable-memory-profiling=YES >> >> and verified that it's configured correctly by examining the output. I >> then >> rebuild R: >> >> % make >> >> Then I fire up R and run a script, using Rprof with the memory-profiling >> switch set to TRUE: >> >> Rprof("output", memory.profiling=TRUE); >> # a bunch of R code >> Rprof(NULL); >> > > > Wen I do > > summaryRprof(memory="both") > > I see an additional column ... > > but since you have not said what you tried exactly, we cannot help very > much. > > Uwe Ligges > > > > When I examine the output, however, using either R CMD Rprof from the >> shell, >> or summaryRprof from within R, the output I see is identical to theoutput>> I >> got when I ran R BEFORE I recompiled with memory profiling enabled. >> >> Anyone see something that I'm missing? >> >> Thanks, Patrick >> >> [[alternative HTML version deleted]] >> >> ______________________________________________ >> R-help at r-project.org mailing list >> https://stat.ethz.ch/mailman/listinfo/r-help >> PLEASE do read the posting guide >> http://www.R-project.org/posting-guide.html >> and provide commented, minimal, self-contained, reproducible code. >> >[[alternative HTML version deleted]] ------------------------------ Message: 58 Date: Mon, 22 Nov 2010 03:14:17 -0800 (PST) From: romzero <romzero at yahoo.it> To: r-help at r-project.org Subject: [R] Some questione about plot Message-ID: <1290424457967-3053430.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii Q1: How can i draw LSD (Least significant difference) on a plot? Like this... http://r.789695.n4.nabble.com/file/n3053430/LSD.jpg Q2: How can i draw the axis secondary scale? Thanks for help. -- View this message in context: http://r.789695.n4.nabble.com/Some-questione-about-plot-tp3053430p3053430.ht ml Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 59 Date: Mon, 22 Nov 2010 08:13:32 -0800 (PST) From: shubha <shuba.pandit at gmail.com> To: r-help at r-project.org Subject: [R] how do remove those predictor which have p value greater than 0.05 in GLM? Message-ID: <1290442412855-3053921.post at n4.nabble.com> Content-Type: text/plain; charset=UTF-8 Hi R user, I am a kind of an intermediate user of R. Now I am using GLM model (library MASS, VEGUS). I used a backward stepwise logistic regression, but i got a problem in removing those predictors which are above 0.05. I don't want to include those variables which were above 0.05 in final backward stepwise logetsic regression model. for example: first I run the model, "name<-glm(dep~env1+env2..., family= binomial, data=new)" after that, I did stepwise for name name.step<-step(name, direction="backward") here, I still got those variables which were not significant, for example: secchi was not significant (see below example), but still it was in the model. how can I remove those variables which are not significant in forward/backward stepwise?. another question, when I wrote direction="backward", I got the results same as in the process of "forward". It is really strange. why is it same results for backward and forward. I checked in other two statistical software (Statistica and SYSTAT), they provided a correct results, I think. But, I need to use R for further analysis, therefore I need to fix the problem. I am spending so much time to figure it out, but I could not. could you please give your suggestions. It would be really a great help. please see the example of retaining predictors which have p value is greater that 0.05 after stepwise logistic regression. Thank Shubha Pandit, PhD University of Windsor Windsor, ON, Canada ===> summary(step.glm.int.ag1)Call: glm(formula = ag1less ~ GEARTEMP + DOGEAR + GEARDEPTH + SECCHI + GEARTEMP:SECCHI + DOGEAR:SECCHI + GEARTEMP:DOGEAR + GEARTEMP:GEARDEPTH + DOGEAR:GEARDEPTH, family = binomial, data = training) Deviance Residuals: Min 1Q Median 3Q Max -2.1983 -0.8272 -0.4677 0.8014 2.6502 Coefficients: Estimate Std. Error z value Pr(>|z|) (Intercept) 3.231623 1.846593 1.750 0.080110 . GEARTEMP -0.004408 0.085254 -0.052 0.958761 DOGEAR -0.732805 0.182285 -4.020 5.82e-05 *** GEARDEPTH -0.249237 0.060825 -4.098 4.17e-05 *** SECCHI 0.311875 0.297594 1.048 0.294645 GEARTEMP:SECCHI -0.080664 0.010079 -8.003 1.21e-15 *** DOGEAR:SECCHI 0.066555 0.022181 3.000 0.002695 ** GEARTEMP:DOGEAR 0.030988 0.008907 3.479 0.000503 *** GEARTEMP:GEARDEPTH 0.008856 0.002122 4.173 3.01e-05 *** DOGEAR:GEARDEPTH 0.006680 0.004483 1.490 0.136151 --- Signif. codes: 0 ?***? 0.001 ?**? 0.01 ?*? 0.05 ?.? 0.1 ? ? 1 (Dispersion parameter for binomial family taken to be 1) Null deviance: 3389.5 on 2751 degrees of freedom Residual devia\ n\ ce: 2720.4 on 2742 degrees of freedom AIC: 2740.4uh Number of Fisher Scoring iterations: 6 =========================>glm.int.ag1<-glm(ag1less~GEARTEMP+DOGEAR+GEARDEPTH+SECCHI+SECCHI*GEARTEMP+SE CCHI*DOGEAR+SECCHI*GEARDEPTH+GEARTEMP*DOGEAR+GEARTEMP*GEARDEPTH+GEARDEPTH*DO GEAR,data=training,> family=binomial) > summary(glm.int.ag1)Call: glm(formula = ag1less ~ GEARTEMP + DOGEAR + GEARDEPTH + SECCHI + SECCHI * GEARTEMP + SECCHI * DOGEAR + SECCHI * GEARDEPTH + GEARTEMP * DOGEAR + GEARTEMP * GEARDEPTH + GEARDEPTH * DOGEAR, family = binomial, data = training) Deviance Residuals: Min 1Q Median 3Q Max -2.1990 -0.8287 -0.4668 0.8055 2.6673 Coefficients: Estimate Std. Error z value Pr(>|z|) (Intercept) 2.909805 1.928375 1.509 0.131314 GEARTEMP 0.005315 0.087159 0.061 0.951379 DOGEAR -0.721864 0.183708 -3.929 8.52e-05 *** GEARDEPTH -0.235961 0.064828 -3.640 0.000273 *** SECCHI 0.391445 0.326542 1.199 0.230622 GEARTEMP:SECCHI -0.082296 0.010437 -7.885 3.14e-15 *** DOGEAR:SECCHI 0.065572 0.022319 2.938 0.003305 ** GEARDEPTH:SECCHI -0.003176 0.005295 -0.600 0.548675 GEARTEMP:DOGEAR 0.030571 0.008961 3.412 0.000646 *** GEARTEMP:GEARDEPTH 0.008692 0.002159 4.027 5.66e-05 *** DOGEAR:GEARDEPTH 0.006544 0.004495 1.456 0.145484 --- Signif. codes: 0 ?***? 0.001 ?**? 0.01 ?*? 0.05 ?.? 0.1 ? ? 1 (Dispersion parameter for binomial family taken to be 1) Null deviance: 3389.5 on 2751 degrees of freedom Residual deviance: 2720.0 on 2741 degrees of freedom AIC: 2742 Number of Fisher Scoring iterations: 6 -- View this message in context: http://r.789695.n4.nabble.com/how-do-remove-those-predictor-which-have-p-val ue-greater-than-0-05-in-GLM-tp3053921p3053921.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 60 Date: Mon, 22 Nov 2010 09:56:37 -0800 (PST) From: wangwallace <talenttree at gmail.com> To: r-help at r-project.org Subject: Re: [R] how to apply sample function to each row of a data frame? Message-ID: <1290448597442-3054117.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii I tried it. it works out perfectly. you save my life. -- View this message in context: http://r.789695.n4.nabble.com/Re-how-to-apply-sample-function-to-each-row-of -a-data-frame-tp3050933p3054117.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 61 Date: Mon, 22 Nov 2010 10:43:15 -0800 (PST) From: tomreilly <tomreilly at autobox.com> To: r-help at r-project.org Subject: Re: [R] arima Message-ID: <1290451395576-3054206.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii Nuncio, No, there is no requirement to subtract the mean. It is required that the residuals are N.I.I.D. (ie constant mean and constant variance). If you have an upward trending series, for example, then the series would need to be "deseasonalized" so that it is constant. There are many many steps to doing this right. Email me at sales at autobox.com to hear more. Tom -- View this message in context: http://r.789695.n4.nabble.com/arima-tp2993543p3054206.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 62 Date: Mon, 22 Nov 2010 11:13:13 -0800 From: Nathan Miller <natemiller77 at gmail.com> To: r-help at r-project.org Subject: [R] Wait for user input with readline() Message-ID: <AANLkTik-U6f+OXv3M=C11+3D7168PwPe1SN6rHcYN5Vv at mail.gmail.com> Content-Type: text/plain Hello, I am trying write a script that includes a prompt for user input using readlines(). I am running into the problem that when I run readlines() as a single line the prompt works perfectly, but when I try to run a block of code which includes the readline function, the script doesn't wait for the user input. I have seen this question posted before when I did a search, but I didn't find an suitable answer. Is there a means of ensuring that the script does not proceed until a value has been entered to readline(). Can I put readline in a function that will wait for input? Are there other options for getting user input that allow require that the script wait for user input? Thanks for your help, Nate [[alternative HTML version deleted]] ------------------------------ Message: 63 Date: Mon, 22 Nov 2010 16:09:40 +0100 From: "Viechtbauer Wolfgang (STAT)" <wolfgang.viechtbauer at maastrichtuniversity.nl> To: Jalla <zazzala at googlemail.com>, "r-help at r-project.org" <r-help at r-project.org> Subject: Re: [R] txtProgressBar strange behavior in R 2.12.0 Message-ID: <077E31A57DA26E46AB0D493C9966AC730BE0812678 at UM-MAIL4112.unimaas.nl> Content-Type: text/plain; charset="utf-8" I believe nobody has responded to far, so maybe this is not a wide-spread issue. However, I have also encountered this since upgrading to R 2.12.0 (Windows 7, 64-bit). In my simulations where I use txtProgressBar(), the problem usually disappears after the bar has progressed to a certain amount, but it's quite strange nonetheless. The characters that appear are gibberish and include some Asian symbols. Here is a screenshot: http://www.wvbauer.com/screenshot.jpg sessionInfo(): R version 2.12.0 (2010-10-15) Platform: x86_64-pc-mingw32/x64 (64-bit) locale: [1] LC_COLLATE=English_United States.1252 [2] LC_CTYPE=English_United States.1252 [3] LC_MONETARY=English_United States.1252 [4] LC_NUMERIC=C [5] LC_TIME=English_United States.1252 attached base packages: [1] stats graphics grDevices utils datasets methods base No idea what could be causing this. Best, -- Wolfgang Viechtbauer Department of Psychiatry and Neuropsychology School for Mental Health and Neuroscience Maastricht University, P.O. Box 616 6200 MD Maastricht, The Netherlands Tel: +31 (43) 368-5248 Fax: +31 (43) 368-8689 Web: http://www.wvbauer.com> -----Original Message----- > From: r-help-bounces at r-project.org [mailto:r-help-bounces at r-project.org] > On Behalf Of Jalla > Sent: Saturday, November 20, 2010 23:49 > To: r-help at r-project.org > Subject: [R] txtProgressBar strange behavior in R 2.12.0 > > > Hi, > I am running R 2.12.0 (windows). > > example(txtProgressBar) > > gives me some funny screen output with all kinds of special characters > appearing and disappearing. It's happening on two different mashines since > vs. 2.12.0. Is this a known issue? > > Best, > Jalla > > -- > View this message in context: > http://r.789695.n4.nabble.com/txtProgressBar-strange-behavior-in-R-2-12-0- > tp3051976p3051976.html > Sent from the R help mailing list archive at Nabble.com. > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide http://www.R-project.org/posting- > guide.html > and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 64 Date: Mon, 22 Nov 2010 17:51:10 +0800 From: ?? <xiagao1982 at gmail.com> To: r-help at r-project.org Subject: [R] How to call web service in R Message-ID: <AANLkTintJO3AVsFhBn0eTrPnOK832E1E7GLa2Wsi9Q+a at mail.gmail.com> Content-Type: text/plain Hello everyone, [[elided Yahoo spam]] Gao Xia [[alternative HTML version deleted]] ------------------------------ Message: 65 Date: Mon, 22 Nov 2010 15:02:20 +0000 From: "Ni, Melody Zhifang" <z.ni at imperial.ac.uk> To: "'r-help at r-project.org'" <r-help at r-project.org> Subject: [R] save a regression model that can be used later Message-ID: <BD9258C2D51F8040927E2E5D2DD39C7C2930926AA6 at ICEXM3.ic.ac.uk> Content-Type: text/plain Hi everyone I have a question about how to save a regression model in R and how to retrieve it for making predictions in a new session. To be more specific, I fitted a multilevel logistic regression model using the lmer from the "lme4" package. I then successfully make predictions using fitted(mymodel). Since data are complex (three levels, nested, numerous categorical and continuous data describing types of laparoscopic surgery), the computer takes quite a while to fit the MLM model. I wonder whether it's possible to save the fitted model so that I don't have to fit it again for making predictions every time I start a new R session. I searched the mailing-list archive. Suggestions include using save () to save the model as "mymodel.rda" and then use load(mymodel.rda) into the workspace. I tried without success (in Windows), returning the error message: "Error in object$fitted : $ operator is invalid for atomic vectors" Did I do anything wrong? Any help on this topic is much appreciated BW, Melody -- Dr Melody Ni Imperial College Department of Surgery and Cancer 10th floor, QEQM Building St. Mary's Hospital London W2 1NY Tel/Fax: +44 (0) 20 331 27657/26309 z.ni at imperial.ac.uk<mailto:z.ni at imperial.ac.uk> [[alternative HTML version deleted]] ------------------------------ Message: 66 Date: Mon, 22 Nov 2010 14:24:47 -0500 From: "Ravi Varadhan" <rvaradhan at jhmi.edu> To: <dhacademic at gmail.com>, <r-help at r-project.org> Subject: Re: [R] question about constraint minimization Message-ID: <004f01cb8a7a$ed137670$c73a6350$@edu> Content-Type: text/plain; charset="us-ascii" I do not understand the constraint x1 = x3 = x4. If this is correct, you only have 10 unknown parameters. If you can correctly formulate your problem, you can have a look at the packages "alabama" or "BB". The function `auglag' in "alabama" or the function `spg' in "BB" may be useful. Ravi. ------------------------------------------------------- Ravi Varadhan, Ph.D. Assistant Professor, Division of Geriatric Medicine and Gerontology School of Medicine Johns Hopkins University Ph. (410) 502-2619 email: rvaradhan at jhmi.edu -----Original Message----- From: r-help-bounces at r-project.org [mailto:r-help-bounces at r-project.org] On Behalf Of dhacademic at gmail.com Sent: Monday, November 22, 2010 11:10 AM To: r-help at r-project.org Subject: Re: [R] question about constraint minimization Hi, I have struggled on this "bound optimization with equality constraint" by using optim function for two days, but still fail to prepare a good input. Can anyone help to prepare the input for my specific case? Many thanks. Best, Hao On Sat, Nov 20, 2010 at 3:17 AM, Hans W Borchers [via R] < ml-node+3051338-309339578-202837 at n4.nabble.com<ml-node%2B3051338-309339578-2 02837 at n4.nabble.com>> wrote:> dhacademic <at> gmail.com <dhacademic <at> gmail.com> writes: > > > > > > > Hi, > > > > I am a beginner of R. There is a question about constraint minimization. > A > > function, y=f(x1,x2,x3....x12), needs to be minimized. There are 3 > > requirements for the minimization: > > > > (1) x2+x3+...+x12=1.5 (x1 is excluded); > > (2) x1=x3=x4; > > (3) x1, x3 and x5 are in the range of -1~0, respectively. The rest > variables > > (x2, x4, x6, x7, ...., x12) are in the range of 0~1, respectively. > > > > The "optim" function is used. And part of my input is as follow, where > > "xx1r" represents the x12: > > > > xx1r=1.5-x[2]-x[1]-x[1]-x[3]-x[4]-x[5]-x[6]-x[7]-x[8]-x[9] > > start=rnorm(9) > > up=1:9/1:9*1 > > lo=1:9/1:9*-1 > > out=optim(start,f,lower=lo,upper=up,method="L-BFGS-B",hessian=TRUE, > > control=list(trace=6,maxit=1000)) > > > > There are two problems in this input. the "up" and "lo" only define a > range > > of -1~1 for x1 to x11, which can not meet the requirement (3). In > addition, > > there is not any constraint imposed on x12. I have no idea how tospecify> a > > matrix that can impose different constraints on individual variables ina> > > function. Any suggestion is highly appreciated. > > > > Best, > > Hao > > > > I don't see any direct need for real 'constraint' optimization here, > it is a 'bounded' optimization where you are allowed to use > > lower <- c(-1,0,-1,0,-1,0,0,0,0,0,0,0) > upper <- c( 0,1, 0,0, 0,1,1,1,1,1,1,1) > > Otherwise, your description is confusing: > (1) Did you change f to a new function with 9 variables, eliminating > x3, x4, and x12 ? > (2) x4 (being equal to x1) has to be in [-1, 0] but also in [0, 1]? > (3) If you need to restrict x12 to [0, 1] also, you cannot eliminate it. > Either keep x12 and use an equality constraint, or use inequality > constraints on xxlr. > > Hans Werner > > ______________________________________________ > [hidden email]<http://user/SendEmail.jtp?type=node&node=3051338&i=0>mailing list> https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. > > > ------------------------------ > View message @ >http://r.789695.n4.nabble.com/question-about-constraint-minimization-tp30508 80p3051338.html> > To unsubscribe from question about constraint minimization, clickhere<http://r.789695.n4.nabble.com/template/NamlServlet.jtp?macro=unsubscrib e_by_code&node=3050880&code=ZGhhY2FkZW1pY0BnbWFpbC5jb218MzA1MDg4MHwtNjM2Nzc0 NA==>.> >-- View this message in context: http://r.789695.n4.nabble.com/question-about-constraint-minimization-tp30508 80p3053912.html Sent from the R help mailing list archive at Nabble.com. [[alternative HTML version deleted]] ______________________________________________ R-help at r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. ------------------------------ Message: 67 Date: Mon, 22 Nov 2010 11:27:09 -0800 From: Joshua Wiley <jwiley.psych at gmail.com> To: bogdanno <bodinsoul at gmail.com> Cc: r-help at r-project.org Subject: Re: [R] Is it possible to make a matrix to start at row 0? Message-ID: <AANLkTikfD_yAkHNwdV26dQmqADiMUmHS_AZ4qfrMy70=@mail.gmail.com> Content-Type: text/plain; charset=UTF-8 Hi, You can do it, but it would be very difficult (think reworking all indexing yourself) and you probably should not even try (nothing else that was expecting indexing to work as the R gods intended it to would work once you had done your rework). What has lead you to want to index from 0? If it is some problem you are having, I can almost certainly promise you it will be easier for us to show you how to approach it differently and index from 1 than to change the underlying framework so you can index from 0. Cheers, Josh On Mon, Nov 22, 2010 at 10:23 AM, bogdanno <bodinsoul at gmail.com> wrote:> > I want to make the matrix to be indexed from row (column) 0, not 1 > Can I do that? How? > Thanks > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.-- Joshua Wiley Ph.D. Student, Health Psychology University of California, Los Angeles http://www.joshuawiley.com/ ------------------------------ Message: 68 Date: Mon, 22 Nov 2010 11:37:29 -0800 From: Bert Gunter <gunter.berton at gene.com> To: bogdanno <bodinsoul at gmail.com> Cc: r-help at r-project.org Subject: Re: [R] Is it possible to make a matrix to start at row 0? Message-ID: <AANLkTinwygygCqK3LNRKm4JrpVOmMek5ecNwjMceL-tA at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Eh??? Why would you want to do that?? (R isn't C). So the simple answer is: you can't. The other answer is, well of course you sort of can via, e.g. for(i in 0:9) { z <- myMatrix[i+1,] ... } But as Josh said, I think this falls into the class of "You are just asking for trouble, so don't do it." Cheers, Bert On Mon, Nov 22, 2010 at 10:23 AM, bogdanno <bodinsoul at gmail.com> wrote:> I want to make the matrix to be indexed from row (column) 0, not 1 > Can I do that? How? > Thanks > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >-- Bert Gunter Genentech Nonclinical Biostatistics ------------------------------ Message: 69 Date: Mon, 22 Nov 2010 14:38:18 -0500 From: Harlan Harris <harlan at harris.name> To: Tal Galili <tal.galili at gmail.com>, Farrel Buchinsky <fjbuch at gmail.com> Cc: r-help at r-project.org, Duncan Temple Lang <duncan at wald.ucdavis.edu> Subject: Re: [R] RGoogleDocs stopped working Message-ID: <AANLkTin=c2Ashu8SQtgV=9A6__MCmyFxwQewMpzoUAML at mail.gmail.com> Content-Type: text/plain No joy for me. :( I'd had version 0.4-1 installed previously, and re-pulling that URL and reinstalling, plus setting RCurlOptions as specified, do not help for me. Exactly the same behavior. It doesn't matter whether I call getGoogleAuth directly or let getGoogleDocsConnection do it for me. -Harlan On Fri, Nov 19, 2010 at 10:27 PM, Farrel Buchinsky <fjbuch at gmail.com> wrote:> Harlan and Tal have had problems. I had lots too. I spent hours getting it > to work. Terrible process to go through but RGoogleDocs is so useful that[[elided Yahoo spam]]> > My problems were overcome when > > 1. I used the latest zip file by Duncan Temple Lang see below > 2. I inserted an options line that loosened the ssl security - do not > know if that was a good thing or not but it got it to work > > Duncan said: > "I have put an updated version of the source of the package with > these changes. It is available from > http://www.omegahat.org/RGoogleDocs/RGoogleDocs_0.4-1.tar.gz > There is a binary for Windows in > http://www.omegahat.org/RGoogleDocs/RGoogleDocs_0.4-1.zip > > Here is my script that works. Does yours look like this? > > library(RGoogleDocs) > packageDescription("RGoogleDocs")> ps <-readline(prompt="get the password in ") > options(RCurlOptions = list(capath = system.file("CurlSSL", "cacert.pem", > package = "RCurl"), ssl.verifypeer = FALSE)) > sheets.con = getGoogleDocsConnection(getGoogleAuth("fjbuch at gmail.com", ps, > service ="wise")) > ts2=getWorksheets("OnCall",sheets.con) #OnCall is just the name of a > spreadsheet > names(ts2) > y2005<-sheetAsMatrix(ts2$y2005,header=TRUE, as.data.frame=TRUE, trim=TRUE) > > Finally, I am willing to offer you a TeamViewer session where we can take > control of one another's computers and see if the problem is code or the > installation. I warn you that I am neither a programmer nor a developer, > just a very enthusiastic RGoogleDocs user who probably perseveres morethan> is good for him. > > Farrel Buchinsky > >On Fri, Nov 19, 2010 at 1:35 PM, Tal Galili <tal.galili at gmail.com> wrote:> I second Harlan's call. > > > ----------------Contact > Details:------------------------------------------------------- > Contact me: Tal.Galili at gmail.com | 972-52-7275845 > Read me: www.talgalili.com (Hebrew) | www.biostatistics.co.il (Hebrew) | > www.r-statistics.com (English) > >---------------------------------------------------------------------------- ------------------> > > > > On Fri, Nov 19, 2010 at 5:00 PM, Harlan Harris <harlan at harris.name> wrote: > >> Any new thoughts on this? I really want to get this working again! Is >> there >> someone else that can help or somewhere else I should be asking? >> >> Thanks! >> >> -Harlan >> >> >> On Wed, Nov 17, 2010 at 10:16 AM, Harlan Harris <harlan at harris.name> >> wrote: >> >> > Following up again. I found on the forums for the Google Apps API this >> > thread that seems to be about a similar issue: >> > >>http://www.google.com/support/forum/p/apps-apis/thread?tid=1c22cb44eb5cbba6& hl=en&search_impression_id=ab161b010ecf8803%3A12c5a65ce83&search_source=rela ted_question>> > >> > It's using Java and is rather over my head, but it seems to suggestthat>> > something related to the content type might be wrong? Does this offer >> any >> > suggestions on how to fix my use of RGoogleDocs? >> > >> > Thanks, >> > >> > -Harlan >> > >> > >> > On Mon, Nov 15, 2010 at 12:16 PM, Harlan Harris <harlan at harris.name >> >wrote: >> > >> >> Thanks, Duncan. Finally getting a chance to follow up on this... >> >> >> >> I tried again, changing and resetting my password, and trying to >> specify >> >> my login and password manually in the getGoogleDocsConnection argument >> list. >> >> I also tried removing either or both of the service and error options. >> No >> >> luck in any case. I also tried a different Google account, also withno>> >> luck. >> >> >> >> I've also tried tweaking the URL being generated by the code, and in >> all >> >> cases, I get a 403: Forbidden error with content >> "Error=BadAuthentication". >> >> >> >> I don't really know enough about how authentication is supposed towork>> to >> >> get much farther. Can you help? Should I try the Google API forum >> instead? >> >> >> >> -Harlan >> >> >> >> >> >> >> >>> From: Duncan Temple Lang <duncan at wald.ucdavis.edu> >> >>> To: r-help at r-project.org >> >>> Date: Wed, 10 Nov 2010 10:33:47 -0800 >> >>> Subject: Re: [R] RGoogleDocs stopped working >> >>> >> >>> Hi Harlan >> >>> >> >>> I just tried to connect to Google Docs and I had ostensibly the same >> >>> problem. >> >>> However, the password was actually different from what I had >> specified. >> >>> After resetting it with GoogleDocs, the getGoogleDocsConnection() >> worked >> >>> fine. So I don't doubt that the login and password are correct, but >> >>> you might just try it again to ensure there are no typos. >> >>> The other thing to look at is the values for Email and Passwd >> >>> sent in the URL, i.e. the string in url in your debugging >> >>> below. (Thanks for that by the way). If either has specialcharacters,>> >>> e.g. &, it is imperative that they are escaped correctly, i.e. >> converted >> >>> to %24. This should happen and nothing should have changed, but itis>> >>> worth verifying. >> >>> >> >>> So things still seem to work for me. It is a data point, but not one >> >>> that gives you much of a clue as to what is wrong on your machine. >> >>> >> >>> D. >> >>> >> >> >> >> On Wed, Nov 10, 2010 at 10:36 AM, Harlan Harris <harlan at harris.name >> >wrote: >> >> >> >>> Hello, >> >>> >> >>> Some code using RGoogleDocs, which had been working smoothly sincethe>> >>> summer, just stopped working. I know that it worked on November 3rd, >> but it >> >>> doesn't work today. I've confirmed that the login and password still >> work >> >>> when I log in manually. I've confirmed that the URL gives the same >> error >> >>> when I paste it into Firefox. I don't know enough about this web >> service to >> >>> figure out the problem myself, alas... >> >>> >> >>> Here's the error and other info (login/password omitted): >> >>> >> >>> > ss.con <- getGoogleDocsConnection(login=gd.login, >> password=gd.password, >> >>> service='wise', error=FALSE) >> >>> Error: Forbidden >> >>> >> >>> Enter a frame number, or 0 to exit >> >>> >> >>> 1: getGoogleDocsConnection(login = gd.login, password = gd.password, >> >>> service = "wise", error = FALSE) >> >>> 2: getGoogleAuth(..., error = error) >> >>> 3: getForm("https://www.google.com/accounts/ClientLogin", accountType >> >> >>> "HOSTED_OR_GOOGLE", Email = login, Passw >> >>> 4: getURLContent(uri, .opts = .opts, .encoding = .encoding, binary >> >>> binary, curl = curl) >> >>> 5: stop.if.HTTP.error(http.header) >> >>> >> >>> Selection: 4 >> >>> Called from: eval(expr, envir, enclos) >> >>> Browse[1]> http.header >> >>> Content-Type >> >>> Cache-control Pragma >> >>> "text/plain" "no-cache, >> >>> no-store" "no-cache" >> >>> Expires Date >> >>> X-Content-Type-Options >> >>> "Mon, 01-Jan-1990 00:00:00 GMT" "Wed, 10 Nov 2010 15:24:39 >> >>> GMT" "nosniff" >> >>> X-XSS-Protection >> >>> Content-Length Server >> >>> "1; mode=block" >> >>> "24" "GSE" >> >>> status statusMessage >> >>> "403" "Forbidden\r\n" >> >>> Browse[1]> url >> >>> [1] " >> >>> >>https://www.google.com/accounts/ClientLogin?accountType=HOSTED%5FOR%5FGOOGLE &Email=***&Passwd=***&service=wise&source=R%2DGoogleDocs%2D0%2E1>> >>> " >> >>> Browse[1]> .opts >> >>> $ssl.verifypeer >> >>> [1] FALSE >> >>> >> >>> >> >>> > R.Version() >> >>> $platform >> >>> [1] "i386-apple-darwin9.8.0" >> >>> >> >>> $arch >> >>> [1] "i386" >> >>> >> >>> $os >> >>> [1] "darwin9.8.0" >> >>> >> >>> $system >> >>> [1] "i386, darwin9.8.0" >> >>> >> >>> $status >> >>> [1] "" >> >>> >> >>> $major >> >>> [1] "2" >> >>> >> >>> $minor >> >>> [1] "10.1" >> >>> >> >>> $year >> >>> [1] "2009" >> >>> >> >>> $month >> >>> [1] "12" >> >>> >> >>> $day >> >>> [1] "14" >> >>> >> >>> $`svn rev` >> >>> [1] "50720" >> >>> >> >>> $language >> >>> [1] "R" >> >>> >> >>> $version.string >> >>> [1] "R version 2.10.1 (2009-12-14)" >> >>> >> >>> >> >>> > installed.packages()[c('RCurl', 'RGoogleDocs'), ] >> >>> Package >> >>> LibPath Version Priority >> Bundle >> >>> Contains >> >>> RCurl "RCurl" >> >>> "/Users/hharris/Library/R/2.10/library" "1.4-3" NA >> NA >> >>> NA >> >>> RGoogleDocs "RGoogleDocs" >> >>> "/Library/Frameworks/R.framework/Resources/library" "0.4-1" NA >> NA >> >>> NA >> >>> Depends Imports LinkingTo >> >>> Suggests Enhances OS_type License Built >> >>> RCurl "R (>= 2.7.0), methods, bitops" NA NA >> >>> "Rcompression" NA NA "BSD" "2.10.1" >> >>> RGoogleDocs "RCurl, XML, methods" NA NA >> >>> NA NA NA "BSD" "2.10.1" >> >>> >> >>>[[elided Yahoo spam]]>> >>> >> >>> -Harlan >> >>> >> >>> >> >> >> > >> >> [[alternative HTML version deleted]] >> >> ______________________________________________ >> R-help at r-project.org mailing list >> https://stat.ethz.ch/mailman/listinfo/r-help >> PLEASE do read the posting guide >> http://www.R-project.org/posting-guide.html >> and provide commented, minimal, self-contained, reproducible code. >> > >[[alternative HTML version deleted]] ------------------------------ Message: 70 Date: Mon, 22 Nov 2010 11:49:43 -0800 From: Peter Ehlers <ehlers at ucalgary.ca> To: Lucia Ca?as <lucia.canas at co.ieo.es> Cc: "r-help at r-project.org" <r-help at r-project.org> Subject: Re: [R] sm.ancova graphic Message-ID: <4CEAC957.8040505 at ucalgary.ca> Content-Type: text/plain; charset=ISO-8859-1; format=flowed On 2010-11-22 09:47, Lucia Ca?as wrote:> > Hi R-Users, > > I am working with sm.ancova (in the package sm) and I have two problemswith the graph, which is automatically generated when sm.ancova() is run.> > 1-Besides of the fitted lines, the observed data appeared automatically inthe graph. I prefer that only fitted lines appear. I check the sm.options, but I could not find the way that the observed data do not appear in the graph.> > 2-I would like to change the size of the numbers in the axis. Again, Icheck the sm.options, but I could not find the correct way.>Your second request is easy: just issue a par(cex.axis = ....) call before the sm.ancova call. For your first problem, I can't see a simple way; this seems to be hard-coded in the function. But the function is easy to modify. Just look for a couple of short loops containing this: text(rawdata$ (It's best to work with the source code in the *.tar.gz file.) Remove or comment out those loops and save the function as, say, my.sm.ancova. You'll also have to set the environment of my.sm.ancova to that of sm.ancova. Seems to me that it might be worth suggesting this and the ability to fiddle with graphics parameters to the maintainer of sm. Peter Ehlers> > > Thank you in advance, > > Luc?a > > [[alternative HTML version deleted]] >------------------------------ Message: 71 Date: Mon, 22 Nov 2010 11:55:10 -0800 From: Joshua Wiley <jwiley.psych at gmail.com> To: Nathan Miller <natemiller77 at gmail.com> Cc: r-help at r-project.org Subject: Re: [R] Wait for user input with readline() Message-ID: <AANLkTi=w+oQ19cinTY8fJ6NtGn=zwi8pNzxT11tOjO=- at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 Hi Nate, There may be better ways, but on the couple instances I've wanted to wait for keyboard input I used this type of paradigm: foo <- function() { x <- 1:10 y <- rnorm(10) input <- NA while(!isTRUE(input == "1") && !isTRUE(input == "2")) { cat("Please type '1' if you want the first variable on the x axis and '2' if you want the second.", fill = TRUE) input <- scan("", what = "character") if(input == "1") { plot(x, y) } else if (input == "2") { plot(y, x) } else {cat("Sorry, I didn't catch that", fill = TRUE)} } } Perhaps it will be of some use to you. Best regards, Josh On Mon, Nov 22, 2010 at 11:13 AM, Nathan Miller <natemiller77 at gmail.com> wrote:> Hello, > > I am trying write a script that includes a prompt for user input using > readlines(). I am running into the problem that when I run readlines() asa> single line the prompt works perfectly, but when I try to run a block of > code which includes the readline function, the script doesn't wait for the > user input. I have seen this question posted before when I did a search,but> I didn't find an suitable answer. Is there a means of ensuring that the > script does not proceed until a value has been entered to readline(). CanI> put readline in a function that will wait for input? > > Are there other options for getting user input that allow require that the > script wait for user input? > > Thanks for your help, > > Nate > > ? ? ? ?[[alternative HTML version deleted]] > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >-- Joshua Wiley Ph.D. Student, Health Psychology University of California, Los Angeles http://www.joshuawiley.com/ ------------------------------ Message: 72 Date: Mon, 22 Nov 2010 14:56:39 -0500 From: Gabor Grothendieck <ggrothendieck at gmail.com> To: Manta <mantino84 at libero.it> Cc: r-help at r-project.org Subject: Re: [R] Ordeing Zoo object Message-ID: <AANLkTikKFhcP7r+EWekDXqx=C06xr+TDp9e+7zbpSQ=S at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 On Mon, Nov 22, 2010 at 1:35 PM, Manta <mantino84 at libero.it> wrote:> > And how about if I want to order the series from the smallest to thelargest> value, keeping the date index in order to see when the values were > predominantly negative etc... >If you just want to look at it in this order then: as.data.frame(dat)[length(dat):1,,drop = FALSE] -- Statistics & Software Consulting GKX Group, GKX Associates Inc. tel: 1-877-GKX-GROUP email: ggrothendieck at gmail.com ------------------------------ Message: 73 Date: Mon, 22 Nov 2010 19:59:23 +0000 (UTC) From: Ben Bolker <bbolker at gmail.com> To: r-help at stat.math.ethz.ch Subject: Re: [R] Is it possible to make a matrix to start at row 0? Message-ID: <loom.20101122T204655-678 at post.gmane.org> Content-Type: text/plain; charset=us-ascii Bert Gunter <gunter.berton <at> gene.com> writes:> > Eh??? Why would you want to do that?? (R isn't C). > > So the simple answer is: you can't. > > The other answer is, well of course you sort of can via, e.g. > > for(i in 0:9) { > z <- myMatrix[i+1,] > ... > } > > But as Josh said, I think this falls into the class of "You are just > asking for trouble, so don't do it." > > Cheers, > BertBut if you still want to after all those warnings, you can ... see the "Oarray" package, where the first letter of the package name is a capital letter "oh" (O), not a zero (0). library("fortunes"); fortune("Yoda") There ought also to be a clever fortune() expressing the sentiment that you may eventually find (weeks, months, or years later) that changing the way you solve your problem to go with R's flow would have been easier than implementing a solution that works around the flow (examples abound: <<-, zero-based arrays, eval(parse()), storing names of variables as character vectors and using get() [FAQ 7.21], etc, etc, etc ...) ------------------------------ Message: 74 Date: Mon, 22 Nov 2010 15:04:01 -0500 From: David Winsemius <dwinsemius at comcast.net> To: "Ni, Melody Zhifang" <z.ni at imperial.ac.uk> Cc: "'r-help at r-project.org'" <r-help at r-project.org> Subject: Re: [R] save a regression model that can be used later Message-ID: <EA74D6E5-DAF0-407E-8451-7F23BD7F84BC at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Nov 22, 2010, at 10:02 AM, Ni, Melody Zhifang wrote:> Hi everyone > > I have a question about how to save a regression model in R and how > to retrieve it for making predictions in a new session. > > To be more specific, I fitted a multilevel logistic regression model > using the lmer from the "lme4" package. I then successfully make > predictions using fitted(mymodel). > > Since data are complex (three levels, nested, numerous categorical > and continuous data describing types of laparoscopic surgery), the > computer takes quite a while to fit the MLM model. I wonder whether > it's possible to save the fitted model so that I don't have to fit > it again for making predictions every time I start a new R session. > > I searched the mailing-list archive. Suggestions include using save > () to save the model as "mymodel.rda" and then use load(mymodel.rda) > into the workspace. I tried without success (in Windows), returning > the error message: "Error in object$fitted : $ operator is invalid > for atomic vectors"How? ... did you "try" that is. Need code, not vague reports of failure.> > Did I do anything wrong? Any help on this topic is much appreciated > > BW, Melody > > -- > Dr Melody Ni > Imperial College > Department of Surgery and Cancer > 10th floor, QEQM Building-- David Winsemius, MD West Hartford, CT ------------------------------ Message: 75 Date: Mon, 22 Nov 2010 12:10:17 -0800 (PST) From: madr <madrazel at interia.pl> To: r-help at r-project.org Subject: [R] how to round only one column of a matrix ? Message-ID: <1290456617008-3054363.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii round() function affects all values of a matrix, I want only to round column that is called 'y'. -- View this message in context: http://r.789695.n4.nabble.com/how-to-round-only-one-column-of-a-matrix-tp305 4363p3054363.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 76 Date: Mon, 22 Nov 2010 21:10:20 +0100 From: baptiste auguie <baptiste.auguie at googlemail.com> To: r-help at stat.math.ethz.ch Subject: Re: [R] Is it possible to make a matrix to start at row 0? Message-ID: <AANLkTi=UTX7mWNAA_wy5Yb=v6oorXTBG21KFwa+-3OnM at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Apparently He who starts from 0 needn't be called unfortunate, fortune('indexed') baptiste On 22 November 2010 20:59, Ben Bolker <bbolker at gmail.com> wrote:> Bert Gunter <gunter.berton <at> gene.com> writes: > >> >> Eh??? Why would you want to do that?? (R isn't C). >> >> So the simple answer is: you can't. >> >> The other answer is, well of course you sort of can via, e.g. >> >> for(i in 0:9) ?{ >> ? ?z <- myMatrix[i+1,] >> ? ... >> } >> >> But as Josh said, I think this falls into the class of "You are just >> asking for trouble, so don't do it." >> >> Cheers, >> Bert > > ?But if you still want to after all those warnings, you can ... > see the "Oarray" package, where the first letter of the package > name is a capital letter "oh" (O), not a zero (0). > > ?library("fortunes"); fortune("Yoda") > > ?There ought also to be a clever fortune() expressing the sentiment > that you may eventually find (weeks, months, or years later) that > changing the way you solve your problem to go with R's flow would > have been easier than implementing a solution that works around > the flow (examples abound: <<-, zero-based arrays, eval(parse()), > storing names ?of variables as character vectors and using get() > [FAQ 7.21], etc, etc, etc ...) > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >------------------------------ Message: 77 Date: Mon, 22 Nov 2010 12:17:15 -0800 (PST) From: Phil Spector <spector at stat.berkeley.edu> To: madr <madrazel at interia.pl> Cc: r-help at r-project.org Subject: Re: [R] how to round only one column of a matrix ? Message-ID: <alpine.DEB.2.00.1011221216480.14043 at springer.Berkeley.EDU> Content-Type: TEXT/PLAIN; charset=US-ASCII; format=flowed Is this what you're looking for?> mymatrix = matrix(rnorm(15),5,3,dimnames=list(NULL,c('x','y','z'))) > mymatrixx y z [1,] -0.4459162 -2.3936837 -0.7401963 [2,] 0.9886466 -1.3955161 -1.3390314 [3,] -0.2086743 1.7984620 -0.8532579 [4,] 1.0985411 0.9315553 -1.3981632 [5,] 0.5787438 0.1719177 0.2246174> mymatrix[,'y'] = round(mymatrix[,'y']) > mymatrixx y z [1,] -0.4459162 -2 -0.7401963 [2,] 0.9886466 -1 -1.3390314 [3,] -0.2086743 2 -0.8532579 [4,] 1.0985411 1 -1.3981632 [5,] 0.5787438 0 0.2246174 - Phil Spector Statistical Computing Facility Department of Statistics UC Berkeley spector at stat.berkeley.edu On Mon, 22 Nov 2010, madr wrote:> > round() function affects all values of a matrix, I want only to roundcolumn> that is called 'y'. > -- > View this message in context:http://r.789695.n4.nabble.com/how-to-round-only-one-column-of-a-matrix-tp305 4363p3054363.html> Sent from the R help mailing list archive at Nabble.com. > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >------------------------------ Message: 78 Date: Mon, 22 Nov 2010 15:19:27 -0500 From: Steve Lianoglou <mailinglist.honeypot at gmail.com> To: ?? <xiagao1982 at gmail.com> Cc: r-help at r-project.org Subject: Re: [R] How to call web service in R Message-ID: <AANLkTimoV-AyGAzu6NVhJX2N6mwRjxFsbmkKXep4pnx4 at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 Hi, On Mon, Nov 22, 2010 at 4:51 AM, ?? <xiagao1982 at gmail.com> wrote:> Hello everyone, >[[elided Yahoo spam]] Is RCurl what you're looking for? http://cran.r-project.org/web/packages/RCurl/index.html -steve -- Steve Lianoglou Graduate Student: Computational Systems Biology ?| Memorial Sloan-Kettering Cancer Center ?| Weill Medical College of Cornell University Contact Info: http://cbio.mskcc.org/~lianos/contact ------------------------------ Message: 79 Date: Mon, 22 Nov 2010 15:39:58 -0500 From: "Tan, Richard" <RTan at panagora.com> To: <r-help at r-project.org> Subject: [R] aggregate a Date column does not work? Message-ID: <3303FA84CE4F7244B27BE264EC4AE2A71510A385 at panemail.panagora.com> Content-Type: text/plain Hi, I am trying to aggregate max a Date type column but have weird result, how do I fix this?> a <- rbind(+ data.frame(name='Tom', payday=as.Date('1999-01-01')), + data.frame(name='Tom', payday=as.Date('2000-01-01')), + data.frame(name='Pete', payday=as.Date('1998-01-01')), + data.frame(name='Pete', payday=as.Date('1999-01-01')) + )> aname payday 1 Tom 1999-01-01 2 Tom 2000-01-01 3 Pete 1998-01-01 4 Pete 1999-01-01> aggregate(a$payday, list(a$name), max)Group.1 x 1 Tom 10957 2 Pete 10592 Thanks, Richard [[alternative HTML version deleted]] ------------------------------ Message: 80 Date: Mon, 22 Nov 2010 12:41:19 -0800 (PST) From: bmiddle <bmiddle at sandia.gov> To: r-help at r-project.org Subject: [R] R2WinBUGS help Message-ID: <1290458479607-3054411.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii When I use the 'bugs' function from R, WinBUGS runs correctly, but R freezes. The only way to use R after this is to stop calculations (without my file that documents the calculation). However, I want to save the output in R so I can use it in further models. Does anyone know how to fix this problem? -- View this message in context: http://r.789695.n4.nabble.com/R2WinBUGS-help-tp3054411p3054411.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 81 Date: Mon, 22 Nov 2010 12:44:36 -0800 (PST) From: tomreilly <tomreilly at autobox.com> To: r-help at r-project.org Subject: Re: [R] FW: help with time Series regression please Message-ID: <1290458676157-3054417.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii Cathy, How does this model look? [(1-B**4)]Y(T) = 20.767 +[X1(T)][(1-B**4)][(+ 56.1962)] :PULSE 7/ 4 I~P00028test +[X2(T)][(1-B**4)][(+ 74.4301)] :PULSE 9/ 4 I~P00036test +[X3(T)][(1-B**4)][(- 59.9872)] :PULSE 6/ 3 I~P00023test +[X4(T)][(1-B**4)][(- 27.2187)] :PULSE 7/ 1 I~P00025test + [(1- .435B** 1)]**-1 [A(T)] -- View this message in context: http://r.789695.n4.nabble.com/errors-appears-in-my-time-Series-regression-fo mula-tp1016593p3054417.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 82 Date: Mon, 22 Nov 2010 15:48:15 -0500 From: Mike Marchywka <marchywka at hotmail.com> To: <bbolker at gmail.com>, <r-help at stat.math.ethz.ch> Subject: Re: [R] Is it possible to make a matrix to start at row 0? Message-ID: <BLU113-W5447FA07F7418936CF2F5BE3D0 at phx.gbl> Content-Type: text/plain; charset="iso-8859-1" ----------------------------------------> To: r-help at stat.math.ethz.ch > From: bbolker at gmail.com > Date: Mon, 22 Nov 2010 19:59:23 +0000 > Subject: Re: [R] Is it possible to make a matrix to start at row 0? > > Bert Gunter gene.com> writes: > > > > > Eh??? Why would you want to do that?? (R isn't C). > > > > So the simple answer is: you can't.Eh, it is open source for a reason. I wouldrecommend however if you let your zero-base version out in the wild you rename it to R0 or something. Seriosuly 1-based arrays annoy me but I'm not sure what happens if you modify R LOL.> > > > The other answer is, well of course you sort of can via, e.g. > > > > for(i in 0:9) { > > z <- myMatrix[i+1,] > > ... > > } > > > > But as Josh said, I think this falls into the class of "You are just > > asking for trouble, so don't do it." > > > > Cheers, > > Bert > > But if you still want to after all those warnings, you can ... > see the "Oarray" package, where the first letter of the package > name is a capital letter "oh" (O), not a zero (0). > > library("fortunes"); fortune("Yoda") > > There ought also to be a clever fortune() expressing the sentiment > that you may eventually find (weeks, months, or years later) that > changing the way you solve your problem to go with R's flow would > have been easier than implementing a solution that works around > the flow (examples abound: <<-, zero-based arrays, eval(parse()), > storing names of variables as character vectors and using get() > [FAQ 7.21], etc, etc, etc ...) > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 83 Date: Mon, 22 Nov 2010 15:50:33 -0500 From: David Winsemius <dwinsemius at comcast.net> To: "Tan, Richard" <RTan at panagora.com> Cc: r-help at r-project.org Subject: Re: [R] aggregate a Date column does not work? Message-ID: <41B45939-081A-4A75-A55C-F5348722918C at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Nov 22, 2010, at 3:39 PM, Tan, Richard wrote:> Hi, I am trying to aggregate max a Date type column but have weird > result, how do I fix this?In the process of getting max() you coerced the Dates to numeric and now you need to re-coerce them back to Dates ?as.Date as.Date(<your result>) (possibly with an origin it the default "1970-01-01" doesn't get used. -- David.> > > >> a <- rbind( > > + data.frame(name='Tom', payday=as.Date('1999-01-01')), > > + data.frame(name='Tom', payday=as.Date('2000-01-01')), > > + data.frame(name='Pete', payday=as.Date('1998-01-01')), > > + data.frame(name='Pete', payday=as.Date('1999-01-01')) > > + ) > >> a > > name payday > > 1 Tom 1999-01-01 > > 2 Tom 2000-01-01 > > 3 Pete 1998-01-01 > > 4 Pete 1999-01-01 > >> aggregate(a$payday, list(a$name), max) > > Group.1 x > > 1 Tom 10957 > > 2 Pete 10592 > > > > Thanks, > > Richard > > > > > [[alternative HTML version deleted]] > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.David Winsemius, MD West Hartford, CT ------------------------------ Message: 84 Date: Mon, 22 Nov 2010 12:53:41 -0800 (PST) From: madr <madrazel at interia.pl> To: r-help at r-project.org Subject: [R] I need a very specific unique like function and I don't know even how to properly call this Message-ID: <1290459221115-3054427.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii consider this matrix: [,1] [,2] [1,] 3 7 [2,] 6 5 [3,] 7 5 [4,] 3 5 [5,] 7 5 [6,] 5 5 [7,] 8 4 [8,] 2 4 [9,] 7 4 [10,] 0 6 I need to delete all rows where column 2 above and below has the same value, so the effect would be: [,1] [,2] [1,] 3 7 [2,] 6 5 [6,] 5 5 [7,] 8 4 [9,] 7 4 [10,] 0 6 is there a built in function for that kind of operation or I must write one from scratch ? Is there a name for that kind of operation ? -- View this message in context: http://r.789695.n4.nabble.com/I-need-a-very-specific-unique-like-function-an d-I-don-t-know-even-how-to-properly-call-this-tp3054427p3054427.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 85 Date: Mon, 22 Nov 2010 15:54:37 -0500 From: "Tan, Richard" <RTan at panagora.com> To: "David Winsemius" <dwinsemius at comcast.net> Cc: r-help at r-project.org Subject: Re: [R] aggregate a Date column does not work? Message-ID: <3303FA84CE4F7244B27BE264EC4AE2A71510A392 at panemail.panagora.com> Content-Type: text/plain; charset="us-ascii" Thanks, add as.Date('1970-01-01') to the result column works. Richard -----Original Message----- From: David Winsemius [mailto:dwinsemius at comcast.net] Sent: Monday, November 22, 2010 3:51 PM To: Tan, Richard Cc: r-help at r-project.org Subject: Re: [R] aggregate a Date column does not work? On Nov 22, 2010, at 3:39 PM, Tan, Richard wrote:> Hi, I am trying to aggregate max a Date type column but have weird > result, how do I fix this?In the process of getting max() you coerced the Dates to numeric and now you need to re-coerce them back to Dates ?as.Date as.Date(<your result>) (possibly with an origin it the default "1970-01-01" doesn't get used. -- David.> > > >> a <- rbind( > > + data.frame(name='Tom', payday=as.Date('1999-01-01')), > > + data.frame(name='Tom', payday=as.Date('2000-01-01')), > > + data.frame(name='Pete', payday=as.Date('1998-01-01')), > > + data.frame(name='Pete', payday=as.Date('1999-01-01')) > > + ) > >> a > > name payday > > 1 Tom 1999-01-01 > > 2 Tom 2000-01-01 > > 3 Pete 1998-01-01 > > 4 Pete 1999-01-01 > >> aggregate(a$payday, list(a$name), max) > > Group.1 x > > 1 Tom 10957 > > 2 Pete 10592 > > > > Thanks, > > Richard > > > > > [[alternative HTML version deleted]] > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.David Winsemius, MD West Hartford, CT ------------------------------ Message: 86 Date: Mon, 22 Nov 2010 15:55:50 -0500 (EST) From: Xiaoqi Cui <xcui at mtu.edu> To: r-help <r-help at r-project.org> Subject: [R] R package "kernlab" can not be properly loaded Message-ID: <1590317927.1718871290459350354.JavaMail.root at marlbed.merit.edu> Content-Type: text/plain; charset=utf-8 Hi, I tried to load the package "kernlab" under R-v11 and R-v10, however it gave error message: Error in library.dynam(lib, package, package.lib) : shared library 'kernlab' not found In addition: Warning message: package 'kernlab' was built under R version 2.12.0 Error: package/namespace load failed for 'kernlab' Has anybody loaded this successfully before? Thanks, Xiaoqi ------------------------------ Message: 87 Date: Mon, 22 Nov 2010 21:56:24 +0100 From: Erich Neuwirth <erich.neuwirth at univie.ac.at> To: csrabak <crabak at acm.org>, r-help <r-help at r-project.org> Subject: Re: [R] Rexcel Message-ID: <4CEAD8F8.4030305 at univie.ac.at> Content-Type: text/plain; charset=ISO-8859-1 a) RExcel has its own mailing list (as the documentation tell you). Please post RExcel related questions to the mailing list accessible at rcom.univie.ac.at b) For running code at startup, you have to create a worksheet (not a workbook) named RCode in your workbook. On 11/22/2010 7:15 PM, csrabak wrote:> Em 22/11/2010 10:11, Luis Felipe Parra escreveu: >> Hello I am new to RExcel and I would like to run a source code form the >> excel worksheet. I would like to run the following code >> >> source("C:\\Quantil Aplicativos\\Genercauca\\BackwardSelectionNC.r") >> >> from the excel wroksheet. Does anybody know how to do this? >> >> Thank you >> > Felipe, > > Look at the section "Startup" in the RExcel help. In a nutshell, if you > want the code to run immediately at the loading of the spreadsheet, > create a workbook called "RCode" and put your source there. > > Other options are available. See the docs. > > -- > Cesar Rabak > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. >------------------------------ Message: 88 Date: Mon, 22 Nov 2010 15:57:45 -0500 From: David Winsemius <dwinsemius at comcast.net> To: "Tan, Richard" <RTan at panagora.com> Cc: r-help at r-project.org Subject: Re: [R] aggregate a Date column does not work? Message-ID: <60654551-AE43-4C06-9AE7-16B9F968324E at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Nov 22, 2010, at 3:54 PM, Tan, Richard wrote:> Thanks, add as.Date('1970-01-01') to the result column works.But that should make them all the same date in 1970. Since aggregate renames the date column to "x", this should work: as.Date( aggregate(a$payday, list(a$name), max)$x ) [1] "2000-01-01" "1999-01-01"> > Richard > > -----Original Message----- > From: David Winsemius [mailto:dwinsemius at comcast.net] > Sent: Monday, November 22, 2010 3:51 PM > To: Tan, Richard > Cc: r-help at r-project.org > Subject: Re: [R] aggregate a Date column does not work? > > > On Nov 22, 2010, at 3:39 PM, Tan, Richard wrote: > >> Hi, I am trying to aggregate max a Date type column but have weird >> result, how do I fix this? > > In the process of getting max() you coerced the Dates to numeric and > now you need to re-coerce them back to Dates > > ?as.Date > as.Date(<your result>) (possibly with an origin it the default > "1970-01-01" doesn't get used. > > -- > David. >> >> >> >>> a <- rbind( >> >> + data.frame(name='Tom', payday=as.Date('1999-01-01')), >> >> + data.frame(name='Tom', payday=as.Date('2000-01-01')), >> >> + data.frame(name='Pete', payday=as.Date('1998-01-01')), >> >> + data.frame(name='Pete', payday=as.Date('1999-01-01')) >> >> + ) >> >>> a >> >> name payday >> >> 1 Tom 1999-01-01 >> >> 2 Tom 2000-01-01 >> >> 3 Pete 1998-01-01 >> >> 4 Pete 1999-01-01 >> >>> aggregate(a$payday, list(a$name), max) >> >> Group.1 x >> >> 1 Tom 10957 >> >> 2 Pete 10592 >> >> >> >> Thanks, >> >> Richard >> >> >> >> >> [[alternative HTML version deleted]] >> >> ______________________________________________ >> R-help at r-project.org mailing list >> https://stat.ethz.ch/mailman/listinfo/r-help >> PLEASE do read the posting guide > http://www.R-project.org/posting-guide.html >> and provide commented, minimal, self-contained, reproducible code. > > David Winsemius, MD > West Hartford, CT > >David Winsemius, MD West Hartford, CT ------------------------------ Message: 89 Date: Mon, 22 Nov 2010 16:08:35 -0500 From: "Tan, Richard" <RTan at panagora.com> To: "David Winsemius" <dwinsemius at comcast.net> Cc: r-help at r-project.org Subject: Re: [R] aggregate a Date column does not work? Message-ID: <3303FA84CE4F7244B27BE264EC4AE2A71510A39F at panemail.panagora.com> Content-Type: text/plain; charset="us-ascii" Yes, I meant something like one of these: b <- aggregate(a$payday, list(a$name), max) b$x <- as.Date('1970-01-01') + b$x or b$x <- as.Date(b$x, origin='1970-01-01') Thanks. -----Original Message----- From: David Winsemius [mailto:dwinsemius at comcast.net] Sent: Monday, November 22, 2010 3:58 PM To: Tan, Richard Cc: r-help at r-project.org Subject: Re: [R] aggregate a Date column does not work? On Nov 22, 2010, at 3:54 PM, Tan, Richard wrote:> Thanks, add as.Date('1970-01-01') to the result column works.But that should make them all the same date in 1970. Since aggregate renames the date column to "x", this should work: as.Date( aggregate(a$payday, list(a$name), max)$x ) [1] "2000-01-01" "1999-01-01"> > Richard > > -----Original Message----- > From: David Winsemius [mailto:dwinsemius at comcast.net] > Sent: Monday, November 22, 2010 3:51 PM > To: Tan, Richard > Cc: r-help at r-project.org > Subject: Re: [R] aggregate a Date column does not work? > > > On Nov 22, 2010, at 3:39 PM, Tan, Richard wrote: > >> Hi, I am trying to aggregate max a Date type column but have weird >> result, how do I fix this? > > In the process of getting max() you coerced the Dates to numeric and > now you need to re-coerce them back to Dates > > ?as.Date > as.Date(<your result>) (possibly with an origin it the default > "1970-01-01" doesn't get used. > > -- > David. >> >> >> >>> a <- rbind( >> >> + data.frame(name='Tom', payday=as.Date('1999-01-01')), >> >> + data.frame(name='Tom', payday=as.Date('2000-01-01')), >> >> + data.frame(name='Pete', payday=as.Date('1998-01-01')), >> >> + data.frame(name='Pete', payday=as.Date('1999-01-01')) >> >> + ) >> >>> a >> >> name payday >> >> 1 Tom 1999-01-01 >> >> 2 Tom 2000-01-01 >> >> 3 Pete 1998-01-01 >> >> 4 Pete 1999-01-01 >> >>> aggregate(a$payday, list(a$name), max) >> >> Group.1 x >> >> 1 Tom 10957 >> >> 2 Pete 10592 >> >> >> >> Thanks, >> >> Richard >> >> >> >> >> [[alternative HTML version deleted]] >> >> ______________________________________________ >> R-help at r-project.org mailing list >> https://stat.ethz.ch/mailman/listinfo/r-help >> PLEASE do read the posting guide > http://www.R-project.org/posting-guide.html >> and provide commented, minimal, self-contained, reproducible code. > > David Winsemius, MD > West Hartford, CT > >David Winsemius, MD West Hartford, CT ------------------------------ Message: 90 Date: Mon, 22 Nov 2010 16:13:11 -0500 From: Gabor Grothendieck <ggrothendieck at gmail.com> To: "Tan, Richard" <RTan at panagora.com> Cc: r-help at r-project.org Subject: Re: [R] aggregate a Date column does not work? Message-ID: <AANLkTik0iSDu3id5WH3BDzEYPv928D7-YWYCW2yzf=rH at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 On Mon, Nov 22, 2010 at 3:39 PM, Tan, Richard <RTan at panagora.com> wrote:> Hi, I am trying to aggregate max a Date type column but have weird > result, how do I fix this? > >> a <- rbind( > > + data.frame(name='Tom', payday=as.Date('1999-01-01')), > + data.frame(name='Tom', payday=as.Date('2000-01-01')), > + data.frame(name='Pete', payday=as.Date('1998-01-01')), > + data.frame(name='Pete', payday=as.Date('1999-01-01')) > + ) >Since its already sorted try this: a[!duplicated(a$name, fromLast = TRUE), ] Using sqldf also works: library(sqldf) sqldf("select name, max(payday) payday from a group by name order by name") -- Statistics & Software Consulting GKX Group, GKX Associates Inc. tel: 1-877-GKX-GROUP email: ggrothendieck at gmail.com ------------------------------ Message: 91 Date: Mon, 22 Nov 2010 13:13:41 -0800 (PST) From: madr <madrazel at interia.pl> To: r-help at r-project.org Subject: Re: [R] "negative alpha" or custom gradient colors of data dots in... Message-ID: <1290460421895-3054465.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii Well I "hacked" it i maybe not so elegant way: x <- rnorm(100000) y <- rnorm(100000) png('out.png',width=600,height=500,units="px") op <- par(bg='black',fg='gray',col='gray',col.axis='gray',col.lab='gray',col.main'gray',col.sub='gray',mai=c(0,0,0,0), tck = 0.01, mgp = c(0, -1.4, 0), xaxs = "i", yaxs = "i") plot(x,y,ylim=c(-20,20),xlim=c(min(x),max(x)),pch='.',col rgb(1,1,1,1),yaxt="n", ann=FALSE) # here is the solution - plot the same data with low alpha and different color blue is best because is perceived as darkest points(x,y,col = rgb(0,0,1,0.1),pch='.') abline(h=0, lty=3, col="green") abline(v=0, lty=3, col="green") par <- op http://i51.tinypic.com/maxmcl.png [[repost, because subject was deleted and this message won't get into the list]] -- View this message in context: http://r.789695.n4.nabble.com/negative-alpha-or-custom-gradient-colors-of-da ta-dots-in-scatterplot-tp3052394p3054465.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 92 Date: Mon, 22 Nov 2010 13:20:16 -0800 (PST) From: Frank Harrell <f.harrell at vanderbilt.edu> To: r-help at r-project.org Subject: Re: [R] how do remove those predictor which have p value greater than 0.05 in GLM? Message-ID: <1290460816357-3054478.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii What would make you want to delete a variable because P > 0.05? That will invalidate every aspect of statistical inference for the model. Frank ----- Frank Harrell Department of Biostatistics, Vanderbilt University -- View this message in context: http://r.789695.n4.nabble.com/how-do-remove-those-predictor-which-have-p-val ue-greater-than-0-05-in-GLM-tp3053921p3054478.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 93 Date: Mon, 22 Nov 2010 15:29:11 -0600 From: lucia <lucia at thedietdiary.com> To: r-help at r-project.org Subject: [R] Help: Standard errors arima Message-ID: <1E4A2E3E-19B8-4531-AEFB-A57CA477CB75 at thedietdiary.com> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes Hello, I'm an R newbie. I've tried to search, but my search skills don't seem up to finding what I need. (Maybe I don't know the correct terms?) I need the standard errors and not the confidence intervals from an ARIMA fit. I can get fits: > coef(test) ar1 ma1 intercept time(TempVector) - 1900 0.801459585 0.704126549 12.854527065 0.000520366 And confidence intervals: > confint(test) 2.5 % 97.5 % ar1 7.684230e-01 0.834496136 ma1 6.742786e-01 0.733974460 intercept 1.217042e+01 13.538635652 time(TempVector) - 1900 -9.610183e-06 0.001050342 > http://stat.ethz.ch/R-manual/R-devel/library/stats/html/arima.html Are any of these standard errors? > vcov(test) ar1 ma1 intercept time(TempVector) - 1900 ar1 2.841144e-04 -5.343792e-05 1.028710e-05 2.725763e-08 ma1 -5.343792e-05 2.319165e-04 9.990842e-07 -3.103661e-09 intercept 1.028710e-05 9.990842e-07 1.218299e-01 8.969206e-05 time(TempVector) - 1900 2.725763e-08 -3.103661e-09 8.969206e-05 7.311670e-08 Or is there a function that can give me standard errors for the coefficients on AR1, ma, and time? (I don't care about the intercept.) Thanks, Lucia ------------------------------ Message: 94 Date: Mon, 22 Nov 2010 16:37:39 -0500 From: David Winsemius <dwinsemius at comcast.net> To: lucia <lucia at thedietdiary.com> Cc: r-help at r-project.org Subject: Re: [R] Help: Standard errors arima Message-ID: <F1D85621-8D27-46B9-B72F-22E82007414B at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Nov 22, 2010, at 4:29 PM, lucia wrote:> Hello, > I'm an R newbie. I've tried to search, but my search skills don't > seem up to finding what I need. (Maybe I don't know the correct > terms?) > > I need the standard errors and not the confidence intervals from an > ARIMA fit. > > I can get fits: > > > coef(test) > ar1 ma1 > intercept time(TempVector) - 1900 > 0.801459585 0.704126549 > 12.854527065 0.000520366 > > And confidence intervals: > > > confint(test) > 2.5 % 97.5 % > ar1 7.684230e-01 0.834496136 > ma1 6.742786e-01 0.733974460 > intercept 1.217042e+01 13.538635652 > time(TempVector) - 1900 -9.610183e-06 0.001050342 > > > > http://stat.ethz.ch/R-manual/R-devel/library/stats/html/arima.htmlThat page says that there is a vcov function that can extract the variance-covariance matrix, so that should let you make s.e.'s pretty easily. -- David.> Are any of these standard errors? > > > vcov(test) > ar1 ma1 intercept > time(TempVector) - 1900 > ar1 2.841144e-04 -5.343792e-05 > 1.028710e-05 2.725763e-08 > ma1 -5.343792e-05 2.319165e-04 > 9.990842e-07 -3.103661e-09 > intercept 1.028710e-05 9.990842e-07 > 1.218299e-01 8.969206e-05 > time(TempVector) - 1900 2.725763e-08 -3.103661e-09 > 8.969206e-05 7.311670e-08 > > Or is there a function that can give me standard errors for the > coefficients on AR1, ma, and time? (I don't care about the > intercept.) > Thanks, > Lucia > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.David Winsemius, MD West Hartford, CT ------------------------------ Message: 95 Date: Mon, 22 Nov 2010 22:41:28 +0100 From: Uwe Ligges <ligges at statistik.tu-dortmund.de> To: bmiddle <bmiddle at sandia.gov> Cc: r-help at r-project.org Subject: Re: [R] R2WinBUGS help Message-ID: <4CEAE388.3070801 at statistik.tu-dortmund.de> Content-Type: text/plain; charset=ISO-8859-1; format=flowed On 22.11.2010 21:41, bmiddle wrote:> > When I use the 'bugs' function from R, WinBUGS runs correctly, but Rfreezes.> The only way to use R after this is to stop calculations (without my file > that documents the calculation).R "freezes" as long as the WinBUGS process is opened. AFter closing it, R should "respond" again. If you like a more inetractive way of dealing with some BUGS incarnation, try the BRugs package that makes use of OpenBUGS by typing install.packages("BRugs") library("BRugs") ?BRugs ?BRugsFit Uwe Ligges However, I want to save the output in R so> I can use it in further models. Does anyone know how to fix this problem?------------------------------ Message: 96 Date: Mon, 22 Nov 2010 16:41:44 -0500 From: David Winsemius <dwinsemius at comcast.net> To: Xiaoqi Cui <xcui at mtu.edu> Cc: r-help <r-help at r-project.org> Subject: Re: [R] R package "kernlab" can not be properly loaded Message-ID: <A0F114AF-60B6-4672-8C2A-85A7848F42C7 at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Nov 22, 2010, at 3:55 PM, Xiaoqi Cui wrote:> Hi, > > I tried to load the package "kernlab" under R-v11 and R-v10, however > it gave error message: > > Error in library.dynam(lib, package, package.lib) : > shared library 'kernlab' not found > In addition: Warning message: > package 'kernlab' was built under R version 2.12.0 > Error: package/namespace load failed for 'kernlab'The current version of R and all of the packages in CRAN _is_ 2.12.0. load()-ing packages that differ in their major versions is not advisable and not guaranteed to succeed. The Contributed Packages page of CRAN has at the very bottom a link to the package Archive. David Winsemius, MD West Hartford, CT ------------------------------ Message: 97 Date: Mon, 22 Nov 2010 22:45:21 +0100 From: Uwe Ligges <ligges at statistik.tu-dortmund.de> To: Xiaoqi Cui <xcui at mtu.edu> Cc: r-help <r-help at r-project.org> Subject: Re: [R] R package "kernlab" can not be properly loaded Message-ID: <4CEAE471.50802 at statistik.tu-dortmund.de> Content-Type: text/plain; charset=ISO-8859-1; format=flowed On 22.11.2010 21:55, Xiaoqi Cui wrote:> Hi, > > I tried to load the package "kernlab" under R-v11 and R-v10, however itgave error message:> > Error in library.dynam(lib, package, package.lib) : > shared library 'kernlab' not found > In addition: Warning message: > package 'kernlab' was built under R version 2.12.0 > Error: package/namespace load failed for 'kernlab' > > Has anybody loaded this successfully before? Thanks,Yes, if it was built for R-2.11.x we were using R-2.11.x. I guess you are under Windows (unstated) where some infrastructure was changed form R-2.11.x to R-2.12.x and packages compiled for the latter won't work foer the former. hcen please run install.packages("kernlab") in order to get a version that fits to your R or even better upgrade your version of R. Uwe Ligges> Xiaoqi > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 98 Date: Mon, 22 Nov 2010 21:51:40 +0000 From: Ista Zahn <izahn at psych.rochester.edu> To: madr <madrazel at interia.pl> Cc: r-help at r-project.org Subject: Re: [R] I need a very specific unique like function and I don't know even how to properly call this Message-ID: <AANLkTin6Yn1FPKiDZO3-WMKFO48_aqA_QFjHKh_4WC7U at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Here is a method for piecing it together using diff and indexing: dat <- structure(c(3L, 6L, 7L, 3L, 7L, 5L, 8L, 2L, 7L, 0L, 7L, 5L, 5L, 5L, 5L, 5L, 4L, 4L, 4L, 6L), .Dim = c(10L, 2L), .Dimnames = list( NULL, c("V1", "V2"))) diffs <- abs(diff(dat[,2], 1)) # get the difference between each value and the previous value new.dat <- cbind(dat, c(NA, diffs), c(diffs, NA)) # combine the diffs with the original matrix, shifted down (is the next valued the same as the value) and down (is the previous value the same) new.dat <- cbind(new.dat, rowSums(new.dat[,3:4], na.rm=TRUE)) # sum the shifted diffs so that the value is 0 if above and below are the same, and greater than zero if the above and below values are not the same final.dat <- new.dat[new.dat[,5] !=0 ,1:2] # get rid of rows for which the sum of the shifted diffs is not equal to zero. HTH, Ista On Mon, Nov 22, 2010 at 8:53 PM, madr <madrazel at interia.pl> wrote:> > consider this matrix: > > ? ? ?[,1] [,2] > ?[1,] ? ?3 ? 7 > ?[2,] ? ?6 ? 5 > ?[3,] ? ?7 ? 5 > ?[4,] ? ?3 ? 5 > ?[5,] ? ?7 ? 5 > ?[6,] ? ?5 ? 5 > ?[7,] ? ?8 ? 4 > ?[8,] ? ?2 ? 4 > ?[9,] ? ?7 ? 4 > [10,] ? ?0 ? 6 > > I need to delete all rows where column 2 above and below has the samevalue,> so the effect would be: > > ? ? ?[,1] [,2] > ?[1,] ? ?3 ? 7 > ?[2,] ? ?6 ? 5 > ?[6,] ? ?5 ? 5 > ?[7,] ? ?8 ? 4 > ?[9,] ? ?7 ? 4 > [10,] ? ?0 ? 6 > > is there a built in function for that kind of operation or I must writeone> from scratch ? > Is there a name for that kind of operation ? > -- > View this message in context:http://r.789695.n4.nabble.com/I-need-a-very-specific-unique-like-function-an d-I-don-t-know-even-how-to-properly-call-this-tp3054427p3054427.html> Sent from the R help mailing list archive at Nabble.com. > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >-- Ista Zahn Graduate student University of Rochester Department of Clinical and Social Psychology http://yourpsyche.org ------------------------------ Message: 99 Date: Mon, 22 Nov 2010 14:03:43 -0800 (PST) From: bmiddle <bmiddle at sandia.gov> To: r-help at r-project.org Subject: Re: [R] R2WinBUGS help Message-ID: <1290463423081-3054531.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii When I use the 'bugs' function from R, WinBUGS runs correctly, but R freezes.> The only way to use R after this is to stop calculations (without my file > that documents the calculation).R "freezes" as long as the WinBUGS process is opened. AFter closing it, R should "respond" again. If you like a more inetractive way of dealing with some BUGS incarnation, try the BRugs package that makes use of OpenBUGS by typing install.packages("BRugs") library("BRugs") ?BRugs ?BRugsFit Thanks, Uwe. I think it was a problem with Windows. WinBUGS should close automatically and it wasn't. Anyway, it seems that the problem is now fixed. I may try the BRugs package soon (after I submit this project report!). -- View this message in context: http://r.789695.n4.nabble.com/R2WinBUGS-help-tp3054411p3054531.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 100 Date: Mon, 22 Nov 2010 14:10:04 -0800 (PST) From: shubha <shuba.pandit at gmail.com> To: r-help at r-project.org Subject: Re: [R] how do remove those predictor which have p value greater than 0.05 in GLM? Message-ID: <1290463804045-3054540.post at n4.nabble.com> Content-Type: text/plain Thanks for the response, Frank. I am not saying that I want to delete a variables because of p>0.5. But my concern was: I am using backward stepwise logistic regression, it keeps the variables in the final model if the variable significantly contributing in the model. Otherwise, it should not be in the final model. Using other software, they give correct results. But R, did not. I want those variables if p<0.05, otherwise exclude from the model. If you include that variables, it will affect the Log likelihood ratio and AIC. I want to change a P-value criterion <=0.05 in the model. Any suggestions. thanks -- View this message in context: http://r.789695.n4.nabble.com/how-do-remove-those-predictor-which-have-p-val ue-greater-than-0-05-in-GLM-tp3053921p3054540.html Sent from the R help mailing list archive at Nabble.com. [[alternative HTML version deleted]] ------------------------------ Message: 101 Date: Mon, 22 Nov 2010 14:11:13 -0800 From: Dennis Murphy <djmuser at gmail.com> To: lucia <lucia at thedietdiary.com> Cc: r-help at r-project.org Subject: Re: [R] Help: Standard errors arima Message-ID: <AANLkTimNjEECU5Ee_+89vsy8D3JzHc9rZbBPhA7458JL at mail.gmail.com> Content-Type: text/plain Hi: Here's an example (never mind the model fit...or lack of it thereof...) str(AirPassengers) # a built-in R data set # Series is seasonal with increasing trend and increasing variance plot(AirPassengers, type = 'l') # STL decomposition plot(stl(AirPassengers, 'periodic')) # ACF and PACF of differenced series par(mfrow = c(2, 1)) acf(diff(AirPassengers)) pacf(diff(AirPassengers)) par(mfrow = c(1, 1)) # Fit a basic seasonal model: SARIMA(0, 1, 1) x (0, 0, 1): m1 <- arima(AirPassengers, order = c(0, 1, 1), seasonal = list(order = c(0, 0, 1), period = 12)) # Most models in R return lists; arima() is no different: names(m1) [1] "coef" "sigma2" "var.coef" "mask" "loglik" "aic" [7] "arma" "residuals" "call" "series" "code" "n.cond" [13] "model" # var.coef looks promising, so let's extract it: m1$var.coef # As David mentioned, vcov() also works (not just for time series, either): vcov(m1) Both should return the same covariance matrix of the estimated coefficients. The standard errors are the square roots of the diagonal elements: sqrt(diag(m1$var.coef)) sqrt(diag(vcov(m1))) Compare this to the output from arima():> m1Call: arima(x = AirPassengers, order = c(0, 1, 1), seasonal = list(order = c(0, 0, 1), period = 12)) Coefficients: ma1 sma1 0.2263 0.8015 s.e. 0.0805 0.0674 HTH, Dennis On Mon, Nov 22, 2010 at 1:29 PM, lucia <lucia at thedietdiary.com> wrote:> Hello, > I'm an R newbie. I've tried to search, but my search skills don't seem up > to finding what I need. (Maybe I don't know the correct terms?) > > I need the standard errors and not the confidence intervals from an ARIMA > fit. > > I can get fits: > > > coef(test) > ar1 ma1 intercept > time(TempVector) - 1900 > 0.801459585 0.704126549 12.854527065 > 0.000520366 > > And confidence intervals: > > > confint(test) > 2.5 % 97.5 % > ar1 7.684230e-01 0.834496136 > ma1 6.742786e-01 0.733974460 > intercept 1.217042e+01 13.538635652 > time(TempVector) - 1900 -9.610183e-06 0.001050342 > > > > http://stat.ethz.ch/R-manual/R-devel/library/stats/html/arima.html > Are any of these standard errors? > > > vcov(test) > ar1 ma1 intercept > time(TempVector) - 1900 > ar1 2.841144e-04 -5.343792e-05 1.028710e-05 > 2.725763e-08 > ma1 -5.343792e-05 2.319165e-04 9.990842e-07 > -3.103661e-09 > intercept 1.028710e-05 9.990842e-07 1.218299e-01 > 8.969206e-05 > time(TempVector) - 1900 2.725763e-08 -3.103661e-09 8.969206e-05 > 7.311670e-08 > > Or is there a function that can give me standard errors for the > coefficients on AR1, ma, and time? (I don't care about the intercept.) > Thanks, > Lucia > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. >[[alternative HTML version deleted]] ------------------------------ Message: 102 Date: Mon, 22 Nov 2010 14:24:50 -0500 From: "Kenney, Colleen T CTR USA AMC" <colleen.t.kenney at us.army.mil> To: <r-help at r-project.org> Subject: [R] Probit Analysis: Confidence Interval for the LD50 using Fieller's and Heterogeneity (UNCLASSIFIED) Message-ID: <0DE122C88FD3464296144E13610ADF700CC59A5C at APGR010BEC80008.nae.ds.army.mil> Content-Type: text/plain; charset="us-ascii" Classification: UNCLASSIFIED Caveats: NONE A similar question has been posted in the past but never answered. My question is this: for probit analysis, how do you program a 95% confidence interval for the LD50 (or LC50, ec50, etc.), including a heterogeneity factor as written about in "Probit Analysis" by Finney(1971)? The heterogeneity factor comes into play through the chi-squared test for homogeneity and is equal to h=chi^2/(k-2), where k is the number of doses and k-2 are the degrees of freedom. I have done a lot of research on this and really appreciate any help [[elided Yahoo spam]] Classification: UNCLASSIFIED Caveats: NONE ------------------------------ Message: 103 Date: Mon, 22 Nov 2010 12:35:55 -0700 From: Yogesh Tiwari <yogesh.mpi at googlemail.com> To: r-help <r-help at stat.math.ethz.ch> Subject: [R] how to calculate derivative Message-ID: <AANLkTind3+hs91==Te155AvYTBB4o4XULE=mEfWTAFwx at mail.gmail.com> Content-Type: text/plain; charset="iso-8859-1" Dear R Users, I have trend of two time series of CO2 each 10 years of data. One is varying weekly and another is bi-weekly. I want to calculate Growth rate ppmv / year of these CO2 trends. Therefore I want to calculate time derivative ppmv / year. How to do it in R? Here I attached example data file, I would appreciate if any one kindly can help on it. Thanks, Regards, Yogesh -------------- next part -------------- 1993.1218 356.8655 1993.1246 356.8723 1993.1273 356.8792 1993.13 356.886 1993.1328 356.8929 1993.1355 356.8998 1993.1383 356.9067 1993.141 356.9136 1993.1437 356.9206 1993.1465 356.9275 1993.1492 356.9345 1993.152 356.9415 1993.1547 356.9485 1993.1574 356.9555 1993.1602 356.9625 1993.1629 356.9695 1993.1656 356.9766 1993.1684 356.9836 1993.1711 356.9907 1993.1739 356.9978 1993.1766 357.0049 1993.1793 357.012 1993.1821 357.0191 1993.1848 357.0263 1993.1875 357.0334 1993.1903 357.0406 1993.193 357.0478 1993.1958 357.0549 1993.1985 357.0621 1993.2012 357.0694 1993.204 357.0766 1993.2067 357.0838 1993.2094 357.0911 1993.2122 357.0983 1993.2149 357.1056 1993.2177 357.1129 1993.2204 357.1202 1993.2231 357.1275 1993.2259 357.1349 1993.2286 357.1422 1993.2313 357.1495 1993.2341 357.1569 1993.2368 357.1643 1993.2396 357.1717 1993.2423 357.1791 1993.245 357.1865 1993.2478 357.1939 1993.2505 357.2013 1993.2533 357.2087 1993.256 357.2162 1993.2587 357.2237 1993.2615 357.2311 1993.2642 357.2386 1993.2669 357.2461 1993.2697 357.2536 1993.2724 357.2611 1993.2752 357.2686 1993.2779 357.2762 1993.2806 357.2837 1993.2834 357.2913 1993.2861 357.2988 1993.2888 357.3064 1993.2916 357.314 1993.2943 357.3216 1993.2971 357.3292 1993.2998 357.3368 1993.3025 357.3444 1993.3053 357.352 1993.308 357.3597 1993.3107 357.3673 1993.3135 357.3747 1993.3162 357.3824 1993.319 357.39 1993.3217 357.3977 1993.3244 357.4054 1993.3272 357.4131 1993.3299 357.4208 1993.3326 357.4285 1993.3354 357.4362 1993.3381 357.4439 1993.3409 357.4516 1993.3436 357.4593 1993.3463 357.4671 1993.3491 357.4748 1993.3518 357.4826 1993.3546 357.4903 1993.3573 357.4981 1993.36 357.5059 1993.3628 357.5136 1993.3655 357.5214 1993.3682 357.5292 1993.371 357.537 1993.3737 357.5448 1993.3765 357.5526 1993.3792 357.5604 1993.3819 357.5683 1993.3847 357.5761 1993.3874 357.5839 1993.3901 357.5918 1993.3929 357.5996 1993.3956 357.6075 1993.3984 357.6153 1993.4011 357.6232 1993.4038 357.631 1993.4066 357.6389 1993.4093 357.6468 1993.412 357.6546 1993.4148 357.6625 1993.4175 357.6702 1993.4203 357.6781 1993.423 357.686 1993.4257 357.6938 1993.4285 357.7017 1993.4312 357.7096 1993.4339 357.7175 1993.4367 357.7254 1993.4394 357.7333 1993.4422 357.7412 1993.4449 357.7491 1993.4476 357.757 1993.4504 357.7649 1993.4531 357.7728 1993.4559 357.7807 1993.4586 357.7886 1993.4613 357.7966 1993.4641 357.8045 1993.4668 357.8124 1993.4695 357.8203 1993.4723 357.8282 1993.475 357.8361 1993.4778 357.8441 1993.4805 357.852 1993.4832 357.8599 1993.486 357.8678 1993.4887 357.8758 1993.4914 357.8837 1993.4942 357.8916 1993.4969 357.8995 1993.4997 357.9074 1993.5024 357.9154 1993.5051 357.9232 1993.5079 357.9311 1993.5106 357.939 1993.5133 357.947 1993.5161 357.9549 1993.5188 357.9628 1993.5216 357.9707 1993.5243 357.9786 1993.527 357.9865 1993.5298 357.9944 1993.5325 358.0023 1993.5352 358.0102 1993.538 358.0181 1993.5407 358.026 1993.5435 358.0339 1993.5462 358.0418 1993.5489 358.0496 1993.5517 358.0575 1993.5544 358.0654 1993.5572 358.0733 1993.5599 358.0811 1993.5626 358.089 1993.5654 358.0969 1993.5681 358.1047 1993.5708 358.1126 1993.5736 358.1204 1993.5763 358.1283 1993.5791 358.1361 1993.5818 358.144 1993.5845 358.1518 1993.5873 358.1596 1993.59 358.1675 1993.5927 358.1753 1993.5955 358.1831 1993.5982 358.1909 1993.601 358.1987 1993.6037 358.2065 1993.6064 358.2142 1993.6092 358.222 1993.6119 358.2298 1993.6146 358.2376 1993.6174 358.2453 1993.6201 358.2531 1993.6229 358.2608 1993.6256 358.2685 1993.6283 358.2763 1993.6311 358.284 1993.6338 358.2917 1993.6366 358.2994 1993.6393 358.3071 1993.642 358.3148 1993.6448 358.3224 1993.6475 358.3301 1993.6502 358.3377 1993.653 358.3454 1993.6557 358.353 1993.6585 358.3607 1993.6612 358.3683 1993.6639 358.3759 1993.6667 358.3835 1993.6694 358.3911 1993.6721 358.3987 1993.6749 358.4062 1993.6776 358.4141 1993.6804 358.4216 1993.6831 358.4292 1993.6858 358.4367 1993.6886 358.4443 1993.6913 358.4518 1993.694 358.4593 1993.6968 358.4668 1993.6995 358.4743 1993.7023 358.4817 1993.705 358.4892 1993.7077 358.4967 1993.7105 358.5041 1993.7132 358.5115 1993.7159 358.5189 1993.7187 358.5263 1993.7214 358.5337 1993.7242 358.5411 1993.7269 358.5484 1993.7296 358.5558 1993.7324 358.5631 1993.7351 358.5704 1993.7379 358.5778 1993.7406 358.585 1993.7433 358.5923 1993.7461 358.5996 1993.7488 358.6068 1993.7515 358.6141 1993.7543 358.6216 1993.757 358.6289 1993.7598 358.6361 1993.7625 358.6433 1993.7652 358.6505 1993.768 358.6576 1993.7707 358.6648 1993.7734 358.6719 1993.7762 358.6791 1993.7789 358.6862 1993.7817 358.6933 1993.7844 358.7003 1993.7871 358.7074 1993.7899 358.7144 1993.7926 358.7215 1993.7953 358.7285 1993.7981 358.7355 1993.8008 358.7424 1993.8036 358.7494 1993.8063 358.7563 1993.809 358.7633 1993.8118 358.7702 1993.8145 358.7771 1993.8172 358.7839 1993.82 358.7908 1993.8227 358.7976 1993.8255 358.8044 1993.8282 358.8117 1993.8309 358.8185 1993.8337 358.8253 1993.8364 358.8321 1993.8392 358.8388 1993.8419 358.8456 1993.8446 358.8523 1993.8474 358.859 1993.8501 358.8657 1993.8528 358.8723 1993.8556 358.879 1993.8583 358.8856 1993.8611 358.8922 1993.8638 358.8988 1993.8665 358.9053 1993.8693 358.9119 1993.872 358.9184 1993.8747 358.9249 1993.8775 358.9314 1993.8802 358.9379 1993.883 358.9443 1993.8857 358.9507 1993.8884 358.9571 1993.8912 358.9635 1993.8939 358.9699 1993.8966 358.9762 1993.8994 358.9825 1993.9021 358.9888 1993.9049 358.9951 1993.9076 359.0018 1993.9103 359.0081 1993.9131 359.0144 1993.9158 359.0206 1993.9185 359.0268 1993.9213 359.033 1993.924 359.0391 1993.9268 359.0452 1993.9295 359.0514 1993.9322 359.0575 1993.935 359.0635 1993.9377 359.0696 1993.9405 359.0756 1993.9432 359.0816 1993.9459 359.0876 1993.9487 359.0935 1993.9514 359.0995 1993.9541 359.1054 1993.9569 359.1112 1993.9596 359.1171 1993.9624 359.1229 1993.9651 359.1288 1993.9678 359.1346 1993.9706 359.1403 1993.9733 359.1461 1993.976 359.1518 1993.9788 359.1575 1993.9815 359.1632 1993.9843 359.1691 1993.987 359.1748 1993.9897 359.1804 1993.9925 359.186 1993.9952 359.1916 1993.9979 359.1971 1994.0007 359.2026 1994.0034 359.2081 1994.0062 359.2136 1994.0089 359.2191 1994.0116 359.2245 1994.0144 359.2299 1994.0171 359.2353 1994.0198 359.2407 1994.0226 359.246 1994.0253 359.2513 1994.0281 359.2566 1994.0308 359.2618 1994.0335 359.2671 1994.0363 359.2723 1994.039 359.2775 1994.0418 359.2826 1994.0445 359.2878 1994.0472 359.2929 1994.05 359.298 1994.0527 359.303 1994.0554 359.3081 1994.0582 359.3131 1994.0609 359.3181 1994.0637 359.323 1994.0664 359.328 1994.0691 359.3329 1994.0719 359.3378 1994.0746 359.3426 1994.0773 359.3475 1994.0801 359.3523 1994.0828 359.3571 1994.0856 359.3618 1994.0883 359.3665 1994.091 359.3713 1994.0938 359.3759 1994.0965 359.3806 1994.0992 359.3852 1994.102 359.3898 1994.1047 359.3944 1994.1075 359.3989 1994.1102 359.4034 1994.1129 359.4079 1994.1157 359.4124 1994.1184 359.4169 1994.1211 359.4213 1994.1239 359.4257 1994.1266 359.4301 1994.1294 359.4344 1994.1321 359.4387 1994.1348 359.443 1994.1376 359.4473 1994.1403 359.4515 1994.1431 359.4558 1994.1458 359.4599 1994.1485 359.4641 1994.1513 359.4683 1994.154 359.4724 1994.1567 359.4765 1994.1595 359.4805 1994.1622 359.4846 1994.165 359.4886 1994.1677 359.4926 1994.1704 359.4966 1994.1732 359.5005 1994.1759 359.5041 1994.1786 359.508 1994.1814 359.5118 1994.1841 359.5156 1994.1869 359.5194 1994.1896 359.5232 1994.1923 359.5269 1994.1951 359.5306 1994.1978 359.5343 1994.2005 359.538 1994.2033 359.5417 1994.206 359.5453 1994.2088 359.5489 1994.2115 359.5524 1994.2142 359.556 1994.217 359.5595 1994.2197 359.563 1994.2225 359.5665 1994.2252 359.5699 1994.2279 359.5733 1994.2307 359.5767 1994.2334 359.5801 1994.2361 359.5835 1994.2389 359.5868 1994.2416 359.5898 1994.2444 359.5931 1994.2471 359.5963 1994.2498 359.5995 1994.2526 359.6027 1994.2553 359.6058 1994.258 359.609 1994.2608 359.6121 1994.2635 359.6152 1994.2663 359.6182 1994.269 359.6213 1994.2717 359.6243 1994.2745 359.6273 1994.2772 359.6303 1994.2799 359.6332 1994.2827 359.6362 1994.2854 359.6391 1994.2882 359.6419 1994.2909 359.6446 1994.2936 359.6475 1994.2964 359.6503 1994.2991 359.653 1994.3018 359.6558 1994.3046 359.6585 1994.3073 359.6612 1994.3101 359.6639 1994.3128 359.6666 1994.3155 359.6692 1994.3183 359.6719 1994.321 359.6745 1994.3238 359.677 1994.3265 359.6796 1994.3292 359.6821 1994.332 359.6846 1994.3347 359.687 1994.3374 359.6895 1994.3402 359.692 1994.3429 359.6944 1994.3457 359.6968 1994.3484 359.6992 1994.3511 359.7015 1994.3539 359.7039 1994.3566 359.7062 1994.3593 359.7085 1994.3621 359.7108 1994.3648 359.713 1994.3676 359.7152 1994.3703 359.7174 1994.373 359.7196 1994.3758 359.7218 1994.3785 359.7239 1994.3812 359.7261 1994.384 359.7282 1994.3867 359.7303 1994.3895 359.7324 1994.3922 359.7344 1994.3949 359.7365 1994.3977 359.7385 1994.4004 359.7405 1994.4031 359.7425 1994.4059 359.7444 1994.4086 359.7464 1994.4114 359.7483 1994.4141 359.7502 1994.4168 359.7521 1994.4196 359.754 1994.4223 359.7558 1994.4251 359.7577 1994.4278 359.7595 1994.4305 359.7613 1994.4333 359.7631 1994.436 359.7649 1994.4387 359.7667 1994.4415 359.7684 1994.4442 359.7701 1994.447 359.7719 1994.4497 359.7736 1994.4524 359.7752 1994.4552 359.7769 1994.4579 359.7786 1994.4606 359.7802 1994.4634 359.7818 1994.4661 359.7834 1994.4689 359.785 1994.4716 359.7866 1994.4743 359.7882 1994.4771 359.7897 1994.4798 359.7913 1994.4825 359.7928 1994.4853 359.7943 1994.488 359.7958 1994.4908 359.7973 1994.4935 359.7988 1994.4962 359.8003 1994.499 359.8017 1994.5017 359.8032 1994.5044 359.8046 1994.5072 359.806 1994.5099 359.8074 1994.5127 359.8088 1994.5154 359.8102 1994.5181 359.8116 1994.5209 359.813 1994.5236 359.8144 1994.5264 359.8157 1994.5291 359.817 1994.5318 359.8184 1994.5346 359.8197 1994.5373 359.821 1994.54 359.8223 1994.5428 359.8236 1994.5455 359.8249 1994.5483 359.8261 1994.551 359.8274 1994.5537 359.8286 1994.5565 359.8299 1994.5592 359.8311 1994.5619 359.8323 1994.5647 359.8336 1994.5674 359.8348 1994.5702 359.836 1994.5729 359.8372 1994.5756 359.8384 1994.5784 359.8398 1994.5811 359.841 1994.5838 359.8422 1994.5866 359.8434 1994.5893 359.8446 1994.5921 359.8457 1994.5948 359.8469 1994.5975 359.8481 1994.6003 359.8492 1994.603 359.8504 1994.6057 359.8515 1994.6085 359.8527 1994.6112 359.8538 1994.614 359.8551 1994.6167 359.8562 1994.6194 359.8574 1994.6222 359.8585 1994.6249 359.8597 1994.6277 359.8608 1994.6304 359.862 1994.6331 359.8631 1994.6359 359.8642 1994.6386 359.8654 1994.6413 359.8665 1994.6441 359.8676 1994.6468 359.8687 1994.6496 359.8698 1994.6523 359.8712 1994.655 359.8723 1994.6578 359.8735 1994.6605 359.8746 1994.6632 359.8758 1994.666 359.8769 1994.6687 359.878 1994.6715 359.8792 1994.6742 359.8803 1994.6769 359.8815 1994.6797 359.8826 1994.6824 359.8838 1994.6851 359.8849 1994.6879 359.8861 1994.6906 359.8875 1994.6934 359.8887 1994.6961 359.8899 1994.6988 359.8911 1994.7016 359.8923 1994.7043 359.8935 1994.707 359.8947 1994.7098 359.8959 1994.7125 359.8971 1994.7153 359.8983 1994.718 359.8995 1994.7207 359.9007 1994.7235 359.9019 1994.7262 359.9032 1994.729 359.9047 1994.7317 359.906 1994.7344 359.9072 1994.7372 359.9085 1994.7399 359.9098 1994.7426 359.9111 1994.7454 359.9124 1994.7481 359.9137 1994.7509 359.9151 1994.7536 359.9164 1994.7563 359.9177 1994.7591 359.9191 1994.7618 359.9204 1994.7645 359.9218 1994.7673 359.9234 1994.77 359.9249 1994.7728 359.9263 1994.7755 359.9277 1994.7782 359.9291 1994.781 359.9306 1994.7837 359.9321 1994.7864 359.9335 1994.7892 359.935 1994.7919 359.9365 1994.7947 359.938 1994.7974 359.9395 1994.8001 359.941 1994.8029 359.9425 1994.8056 359.9441 1994.8084 359.946 1994.8111 359.9476 1994.8138 359.9492 1994.8166 359.9509 1994.8193 359.9525 1994.822 359.9542 1994.8248 359.9558 1994.8275 359.9575 1994.8303 359.9592 1994.833 359.9609 1994.8357 359.9626 1994.8385 359.9644 1994.8412 359.9661 1994.8439 359.9679 1994.8467 359.97 1994.8494 359.9719 1994.8522 359.9737 1994.8549 359.9756 1994.8576 359.9775 1994.8604 359.9794 1994.8631 359.9813 1994.8658 359.9832 1994.8686 359.9852 1994.8713 359.9871 1994.8741 359.9891 1994.8768 359.9911 1994.8795 359.9931 1994.8823 359.9952 1994.885 359.9976 1994.8877 359.9997 1994.8905 360.0018 1994.8932 360.004 1994.896 360.0062 1994.8987 360.0084 1994.9014 360.0106 1994.9042 360.0128 1994.9069 360.015 1994.9097 360.0173 1994.9124 360.0196 1994.9151 360.0219 1994.9179 360.0242 1994.9206 360.0269 1994.9233 360.0293 1994.9261 360.0317 1994.9288 360.0342 1994.9316 360.0367 1994.9343 360.0392 1994.937 360.0417 1994.9398 360.0442 1994.9425 360.0468 1994.9452 360.0494 1994.948 360.052 1994.9507 360.0548 1994.9535 360.0575 1994.9562 360.0602 1994.9589 360.063 1994.9617 360.0657 1994.9644 360.0685 1994.9671 360.0713 1994.9699 360.0741 1994.9726 360.077 1994.9754 360.0799 1994.9781 360.0828 1994.9808 360.0857 1994.9836 360.0886 1994.9863 360.0916 1994.989 360.0946 1994.9918 360.0976 1994.9945 360.1006 1994.9973 360.1043 1995 360.1075 1995.0027 360.1107 1995.0055 360.1139 1995.0082 360.1171 1995.011 360.1204 1995.0137 360.1237 1995.0164 360.127 1995.0192 360.1303 1995.0219 360.1337 1995.0246 360.1371 1995.0274 360.1405 1995.0301 360.1439 1995.0329 360.1474 1995.0356 360.1513 1995.0383 360.1549 1995.0411 360.1585 1995.0438 360.1622 1995.0465 360.1658 1995.0493 360.1695 1995.052 360.1733 1995.0548 360.177 1995.0575 360.1808 1995.0602 360.1846 1995.063 360.1884 1995.0657 360.1923 1995.0684 360.1962 1995.0712 360.2001 1995.0739 360.2045 1995.0767 360.2085 1995.0794 360.2125 1995.0821 360.2166 1995.0849 360.2208 1995.0876 360.2249 1995.0903 360.2291 1995.0931 360.2333 1995.0958 360.2375 1995.0986 360.2418 1995.1013 360.2461 1995.104 360.2504 1995.1068 360.2548 1995.1095 360.2592 1995.1123 360.2636 1995.115 360.268 1995.1177 360.2731 1995.1205 360.2776 1995.1232 360.2822 1995.1259 360.2869 1995.1287 360.2915 1995.1314 360.2962 1995.1342 360.301 1995.1369 360.3057 1995.1396 360.3105 1995.1424 360.3153 1995.1451 360.3202 1995.1478 360.3251 1995.1506 360.33 1995.1533 360.3353 1995.1561 360.3403 1995.1588 360.3454 1995.1615 360.3505 1995.1643 360.3556 1995.167 360.3608 1995.1697 360.366 1995.1725 360.3712 1995.1752 360.3764 1995.178 360.3817 1995.1807 360.387 1995.1834 360.3924 1995.1862 360.3978 1995.1889 360.4032 1995.1916 360.4087 1995.1944 360.4146 1995.1971 360.4201 1995.1999 360.4257 1995.2026 360.4314 1995.2053 360.437 1995.2081 360.4428 1995.2108 360.4485 1995.2136 360.4543 1995.2163 360.4601 1995.219 360.4659 1995.2218 360.4718 1995.2245 360.4777 1995.2272 360.4839 1995.23 360.4899 1995.2327 360.4959 1995.2355 360.502 1995.2382 360.5081 1995.2409 360.5143 1995.2437 360.5204 1995.2464 360.5267 1995.2491 360.5329 1995.2519 360.5392 1995.2546 360.5455 1995.2574 360.5519 1995.2601 360.5582 1995.2628 360.5647 1995.2656 360.5711 1995.2683 360.5776 1995.271 360.5841 1995.2738 360.5907 1995.2765 360.5973 1995.2793 360.6039 1995.282 360.6112 1995.2847 360.6179 1995.2875 360.6247 1995.2902 360.6315 1995.293 360.6384 1995.2957 360.6452 1995.2984 360.6522 1995.3012 360.6591 1995.3039 360.6661 1995.3066 360.6731 1995.3094 360.6802 1995.3121 360.6873 1995.3149 360.6944 1995.3176 360.7016 1995.3203 360.7088 1995.3231 360.7163 1995.3258 360.7236 1995.3285 360.7309 1995.3313 360.7383 1995.334 360.7457 1995.3368 360.7531 1995.3395 360.7606 1995.3422 360.7681 1995.345 360.7756 1995.3477 360.7832 1995.3504 360.7908 1995.3532 360.7985 1995.3559 360.8061 1995.3587 360.8138 1995.3614 360.8216 1995.3641 360.8294 1995.3669 360.8374 1995.3696 360.8453 1995.3723 360.8532 1995.3751 360.8612 1995.3778 360.8691 1995.3806 360.8772 1995.3833 360.8852 1995.386 360.8933 1995.3888 360.9014 1995.3915 360.9095 1995.3943 360.9177 1995.397 360.9259 1995.3997 360.9341 1995.4025 360.9426 1995.4052 360.9509 1995.4079 360.9593 1995.4107 360.9677 1995.4134 360.9761 1995.4162 360.9845 1995.4189 360.993 1995.4216 361.0016 1995.4244 361.0101 1995.4271 361.0187 1995.4298 361.0273 1995.4326 361.036 1995.4353 361.0447 1995.4381 361.0534 1995.4408 361.0623 1995.4435 361.0711 1995.4463 361.0799 1995.449 361.0888 1995.4517 361.0977 1995.4545 361.1066 1995.4572 361.1156 1995.46 361.1245 1995.4627 361.1336 1995.4654 361.1426 1995.4682 361.1517 1995.4709 361.1608 1995.4736 361.1699 1995.4764 361.1791 1995.4791 361.1883 1995.4819 361.1977 1995.4846 361.207 1995.4873 361.2163 1995.4901 361.2256 1995.4928 361.235 1995.4956 361.2444 1995.4983 361.2538 1995.501 361.2633 1995.5038 361.2728 1995.5065 361.2823 1995.5092 361.2918 1995.512 361.3014 1995.5147 361.3111 1995.5175 361.3207 1995.5202 361.3304 1995.5229 361.3401 1995.5257 361.3498 1995.5284 361.3595 1995.5311 361.3693 1995.5339 361.3791 1995.5366 361.389 1995.5394 361.3988 1995.5421 361.4087 1995.5448 361.4186 1995.5476 361.4286 1995.5503 361.4385 1995.553 361.4486 1995.5558 361.4586 1995.5585 361.4687 1995.5613 361.4788 1995.564 361.4889 1995.5667 361.499 1995.5695 361.5092 1995.5722 361.5194 1995.5749 361.5296 1995.5777 361.5398 1995.5804 361.5501 1995.5832 361.5604 1995.5859 361.5707 1995.5886 361.581 1995.5914 361.5914 1995.5941 361.6018 1995.5969 361.6122 1995.5996 361.6226 1995.6023 361.6331 1995.6051 361.6436 1995.6078 361.6541 1995.6105 361.6646 1995.6133 361.6751 1995.616 361.6857 1995.6188 361.6963 1995.6215 361.7069 1995.6242 361.7176 1995.627 361.7282 1995.6297 361.7389 1995.6324 361.7496 1995.6352 361.7603 1995.6379 361.771 1995.6407 361.7818 1995.6434 361.7926 1995.6461 361.8034 1995.6489 361.8142 1995.6516 361.8251 1995.6543 361.8359 1995.6571 361.8468 1995.6598 361.8577 1995.6626 361.8686 1995.6653 361.8795 1995.668 361.8905 1995.6708 361.9014 1995.6735 361.9124 1995.6762 361.9234 1995.679 361.9344 1995.6817 361.9455 1995.6845 361.9565 1995.6872 361.9676 1995.6899 361.9787 1995.6927 361.9898 1995.6954 362.0009 1995.6982 362.012 1995.7009 362.0232 1995.7036 362.0343 1995.7064 362.0455 1995.7091 362.0566 1995.7118 362.0678 1995.7146 362.079 1995.7173 362.0903 1995.7201 362.1015 1995.7228 362.1128 1995.7255 362.124 1995.7283 362.1353 1995.731 362.1466 1995.7337 362.1579 1995.7365 362.1693 1995.7392 362.1806 1995.742 362.1919 1995.7447 362.2033 1995.7474 362.2145 1995.7502 362.2259 1995.7529 362.2372 1995.7556 362.2486 1995.7584 362.26 1995.7611 362.2714 1995.7639 362.2828 1995.7666 362.2943 1995.7693 362.3057 1995.7721 362.3171 1995.7748 362.3286 1995.7775 362.34 1995.7803 362.3515 1995.783 362.363 1995.7858 362.3745 1995.7885 362.3858 1995.7912 362.3972 1995.794 362.4087 1995.7967 362.4202 1995.7995 362.4317 1995.8022 362.4432 1995.8049 362.4547 1995.8077 362.4662 1995.8104 362.4777 1995.8131 362.4892 1995.8159 362.5008 1995.8186 362.5123 1995.8214 362.5238 1995.8241 362.5352 1995.8268 362.5467 1995.8296 362.5582 1995.8323 362.5697 1995.835 362.5813 1995.8378 362.5928 1995.8405 362.6043 1995.8433 362.6159 1995.846 362.6274 1995.8487 362.6389 1995.8515 362.6505 1995.8542 362.662 1995.8569 362.6735 1995.8597 362.6849 1995.8624 362.6964 1995.8652 362.7079 1995.8679 362.7194 1995.8706 362.7309 1995.8734 362.7424 1995.8761 362.7539 1995.8789 362.7654 1995.8816 362.7769 1995.8843 362.7884 1995.8871 362.7999 1995.8898 362.8114 1995.8925 362.8229 1995.8953 362.8344 1995.898 362.8459 1995.9008 362.857 1995.9035 362.8685 1995.9062 362.8799 1995.909 362.8913 1995.9117 362.9028 1995.9144 362.9142 1995.9172 362.9256 1995.9199 362.937 1995.9227 362.9484 1995.9254 362.9598 1995.9281 362.9712 1995.9309 362.9826 1995.9336 362.994 1995.9363 363.0054 1995.9391 363.0164 1995.9418 363.0277 1995.9446 363.039 1995.9473 363.0503 1995.95 363.0616 1995.9528 363.0729 1995.9555 363.0841 1995.9582 363.0954 1995.961 363.1067 1995.9637 363.1179 1995.9665 363.1291 1995.9692 363.1404 1995.9719 363.1516 1995.9747 363.1628 1995.9774 363.174 1995.9802 363.1848 1995.9829 363.1959 1995.9856 363.207 1995.9884 363.2181 1995.9911 363.2291 1995.9938 363.2402 1995.9966 363.2513 1995.9993 363.2623 1996.0021 363.2733 1996.0048 363.2844 1996.0075 363.2954 1996.0103 363.3063 1996.013 363.3173 1996.0157 363.3283 1996.0185 363.3392 1996.0212 363.3502 1996.024 363.3611 1996.0267 363.3714 1996.0294 363.3822 1996.0322 363.393 1996.0349 363.4038 1996.0376 363.4146 1996.0404 363.4253 1996.0431 363.4361 1996.0459 363.4468 1996.0486 363.4575 1996.0513 363.4682 1996.0541 363.4788 1996.0568 363.4895 1996.0595 363.5001 1996.0623 363.5107 1996.065 363.5213 1996.0678 363.5318 1996.0705 363.5424 1996.0732 363.5524 1996.076 363.5628 1996.0787 363.5732 1996.0815 363.5836 1996.0842 363.594 1996.0869 363.6044 1996.0897 363.6147 1996.0924 363.625 1996.0951 363.6353 1996.0979 363.6456 1996.1006 363.6558 1996.1034 363.6661 1996.1061 363.6763 1996.1088 363.6861 1996.1116 363.6963 1996.1143 363.7063 1996.117 363.7164 1996.1198 363.7264 1996.1225 363.7364 1996.1253 363.7464 1996.128 363.7564 1996.1307 363.7663 1996.1335 363.7763 1996.1362 363.7861 1996.1389 363.796 1996.1417 363.8058 1996.1444 363.8156 1996.1472 363.8254 1996.1499 363.8352 1996.1526 363.8449 1996.1554 363.8546 1996.1581 363.8643 1996.1608 363.8739 1996.1636 363.8836 1996.1663 363.8931 1996.1691 363.9027 1996.1718 363.9122 1996.1745 363.9207 1996.1773 363.9301 1996.18 363.9395 1996.1828 363.9488 1996.1855 363.9581 1996.1882 363.9674 1996.191 363.9766 1996.1937 363.9858 1996.1964 363.995 1996.1992 364.0042 1996.2019 364.0133 1996.2047 364.0222 1996.2074 364.0312 1996.2101 364.0402 1996.2129 364.0491 1996.2156 364.058 1996.2183 364.0669 1996.2211 364.0758 1996.2238 364.0846 1996.2266 364.0934 1996.2293 364.1022 1996.232 364.1109 1996.2348 364.1196 1996.2375 364.1283 1996.2402 364.1369 1996.243 364.1455 1996.2457 364.1537 1996.2485 364.1621 1996.2512 364.1706 1996.2539 364.179 1996.2567 364.1874 1996.2594 364.1957 1996.2621 364.204 1996.2649 364.2123 1996.2676 364.2205 1996.2704 364.2287 1996.2731 364.2369 1996.2758 364.245 1996.2786 364.2531 1996.2813 364.2611 1996.2841 364.2688 1996.2868 364.2767 1996.2895 364.2846 1996.2923 364.2925 1996.295 364.3003 1996.2977 364.3081 1996.3005 364.3158 1996.3032 364.3236 1996.306 364.3312 1996.3087 364.3389 1996.3114 364.3465 1996.3142 364.354 1996.3169 364.3615 1996.3196 364.369 1996.3224 364.3765 1996.3251 364.3835 1996.3279 364.3908 1996.3306 364.3981 1996.3333 364.4053 1996.3361 364.4125 1996.3388 364.4197 1996.3415 364.4268 1996.3443 364.4339 1996.347 364.4409 1996.3498 364.4479 1996.3525 364.4549 1996.3552 364.4618 1996.358 364.4685 1996.3607 364.4753 1996.3634 364.4821 1996.3662 364.4888 1996.3689 364.4955 1996.3717 364.5021 1996.3744 364.5087 1996.3771 364.5153 1996.3799 364.5218 1996.3826 364.5283 1996.3854 364.5347 1996.3881 364.5411 1996.3908 364.5475 1996.3936 364.5538 1996.3963 364.5601 1996.399 364.5659 1996.4018 364.5721 1996.4045 364.5782 1996.4073 364.5843 1996.41 364.5903 1996.4127 364.5963 1996.4155 364.6022 1996.4182 364.6081 1996.4209 364.6139 1996.4237 364.6198 1996.4264 364.6255 1996.4292 364.6313 1996.4319 364.6369 1996.4346 364.6426 1996.4374 364.6479 1996.4401 364.6534 1996.4428 364.6589 1996.4456 364.6643 1996.4483 364.6697 1996.4511 364.675 1996.4538 364.6803 1996.4565 364.6856 1996.4593 364.6908 1996.462 364.696 1996.4648 364.7011 1996.4675 364.7062 1996.4702 364.7113 1996.473 364.716 1996.4757 364.7209 1996.4784 364.7258 1996.4812 364.7307 1996.4839 364.7354 1996.4867 364.7402 1996.4894 364.7449 1996.4921 364.7496 1996.4949 364.7542 1996.4976 364.7588 1996.5003 364.7633 1996.5031 364.7678 1996.5058 364.7722 1996.5086 364.7766 1996.5113 364.781 1996.514 364.785 1996.5168 364.7892 1996.5195 364.7934 1996.5222 364.7976 1996.525 364.8017 1996.5277 364.8057 1996.5305 364.8097 1996.5332 364.8137 1996.5359 364.8176 1996.5387 364.8215 1996.5414 364.8254 1996.5441 364.8292 1996.5469 364.8329 1996.5496 364.8364 1996.5524 364.84 1996.5551 364.8436 1996.5578 364.8472 1996.5606 364.8507 1996.5633 364.8542 1996.5661 364.8576 1996.5688 364.8609 1996.5715 364.8643 1996.5743 364.8676 1996.577 364.8708 1996.5797 364.874 1996.5825 364.8772 1996.5852 364.8803 1996.588 364.8834 1996.5907 364.8861 1996.5934 364.8891 1996.5962 364.892 1996.5989 364.8948 1996.6016 364.8977 1996.6044 364.9005 1996.6071 364.9032 1996.6099 364.9059 1996.6126 364.9086 1996.6153 364.9112 1996.6181 364.9137 1996.6208 364.9163 1996.6235 364.9188 1996.6263 364.9212 1996.629 364.9236 1996.6318 364.926 1996.6345 364.9283 1996.6372 364.9306 1996.64 364.9329 1996.6427 364.9351 1996.6454 364.9372 1996.6482 364.9394 1996.6509 364.9414 1996.6537 364.9435 1996.6564 364.9455 1996.6591 364.9475 1996.6619 364.9494 1996.6646 364.9502 1996.6674 364.9519 1996.6701 364.9537 1996.6728 364.9553 1996.6756 364.957 1996.6783 364.9586 1996.681 364.9602 1996.6838 364.9617 1996.6865 364.9632 1996.6893 364.9646 1996.692 364.966 1996.6947 364.9674 1996.6975 364.9687 1996.7002 364.97 1996.7029 364.9713 1996.7057 364.9722 1996.7084 364.9733 1996.7112 364.9744 1996.7139 364.9755 1996.7166 364.9765 1996.7194 364.9775 1996.7221 364.9784 1996.7248 364.9793 1996.7276 364.9802 1996.7303 364.981 1996.7331 364.9818 1996.7358 364.9826 1996.7385 364.9833 1996.7413 364.9837 1996.744 364.9844 1996.7467 364.9849 1996.7495 364.9855 1996.7522 364.986 1996.755 364.9865 1996.7577 364.9869 1996.7604 364.9873 1996.7632 364.9877 1996.7659 364.988 1996.7687 364.9883 1996.7714 364.9886 1996.7741 364.9888 1996.7769 364.989 1996.7796 364.9892 1996.7823 364.9889 1996.7851 364.989 1996.7878 364.989 1996.7906 364.989 1996.7933 364.9889 1996.796 364.9888 1996.7988 364.9887 1996.8015 364.9886 1996.8042 364.9884 1996.807 364.9882 1996.8097 364.9879 1996.8125 364.9877 1996.8152 364.9874 1996.8179 364.987 1996.8207 364.9863 1996.8234 364.9859 1996.8261 364.9854 1996.8289 364.9849 1996.8316 364.9844 1996.8344 364.9838 1996.8371 364.9832 1996.8398 364.9826 1996.8426 364.982 1996.8453 364.9813 1996.848 364.9806 1996.8508 364.9798 1996.8535 364.9791 1996.8563 364.9783 1996.859 364.9771 1996.8617 364.9762 1996.8645 364.9753 1996.8672 364.9744 1996.87 364.9734 1996.8727 364.9724 1996.8754 364.9714 1996.8782 364.9703 1996.8809 364.9692 1996.8836 364.9681 1996.8864 364.967 1996.8891 364.9659 1996.8919 364.9647 1996.8946 364.9635 1996.8973 364.9619 1996.9001 364.9606 1996.9028 364.9593 1996.9055 364.9579 1996.9083 364.9566 1996.911 364.9552 1996.9138 364.9538 1996.9165 364.9523 1996.9192 364.9509 1996.922 364.9494 1996.9247 364.9479 1996.9274 364.9464 1996.9302 364.9448 1996.9329 364.9433 1996.9357 364.9417 1996.9384 364.9401 1996.9411 364.9384 1996.9439 364.9368 1996.9466 364.9351 1996.9493 364.9335 1996.9521 364.9318 1996.9548 364.93 1996.9576 364.9283 1996.9603 364.9266 1996.963 364.9248 1996.9658 364.923 1996.9685 364.9212 1996.9713 364.9194 1996.974 364.9175 1996.9767 364.9157 1996.9795 364.9138 1996.9822 364.9119 1996.9849 364.91 1996.9877 364.9081 1996.9904 364.9062 1996.9932 364.9042 1996.9959 364.9023 1996.9986 364.9003 1997.0014 364.8983 1997.0041 364.8963 1997.0068 364.8943 1997.0096 364.8922 1997.0123 364.8902 1997.0151 364.8882 1997.0178 364.8861 1997.0205 364.8806 1997.0233 364.8783 1997.026 364.8761 1997.0287 364.8738 1997.0315 364.8716 1997.0342 364.8693 1997.037 364.867 1997.0397 364.8647 1997.0424 364.8624 1997.0452 364.8601 1997.0479 364.8578 1997.0507 364.8553 1997.0534 364.8529 1997.0561 364.8505 1997.0589 364.8481 1997.0616 364.8457 1997.0643 364.8433 1997.0671 364.8409 1997.0698 364.8385 1997.0726 364.8361 1997.0753 364.8337 1997.078 364.8312 1997.0808 364.8286 1997.0835 364.8262 1997.0862 364.8237 1997.089 364.8212 1997.0917 364.8187 1997.0945 364.8162 1997.0972 364.8137 1997.0999 364.8113 1997.1027 364.8088 1997.1054 364.8063 1997.1081 364.8038 1997.1109 364.8013 1997.1136 364.7988 1997.1164 364.796 1997.1191 364.7935 1997.1218 364.7909 1997.1246 364.7884 1997.1273 364.7859 1997.13 364.7833 1997.1328 364.7808 1997.1355 364.7783 1997.1383 364.7757 1997.141 364.7732 1997.1437 364.7707 1997.1465 364.7681 1997.1492 364.7656 1997.152 364.7631 1997.1547 364.7603 1997.1574 364.7578 1997.1602 364.7552 1997.1629 364.7527 1997.1656 364.7501 1997.1684 364.7476 1997.1711 364.7451 1997.1739 364.7426 1997.1766 364.7401 1997.1793 364.7375 1997.1821 364.735 1997.1848 364.7324 1997.1875 364.7299 1997.1903 364.7274 1997.193 364.7249 1997.1958 364.7224 1997.1985 364.7199 1997.2012 364.7174 1997.204 364.7149 1997.2067 364.7125 1997.2094 364.71 1997.2122 364.7076 1997.2149 364.7052 1997.2177 364.7028 1997.2204 364.7003 1997.2231 364.698 1997.2259 364.6954 1997.2286 364.693 1997.2313 364.6906 1997.2341 364.6882 1997.2368 364.6858 1997.2396 364.6835 1997.2423 364.6811 1997.245 364.6788 1997.2478 364.6765 1997.2505 364.6742 1997.2533 364.6719 1997.256 364.6697 1997.2587 364.6674 1997.2615 364.6651 1997.2642 364.6628 1997.2669 364.6606 1997.2697 364.6584 1997.2724 364.6562 1997.2752 364.654 1997.2779 364.6519 1997.2806 364.6497 1997.2834 364.6476 1997.2861 364.6455 1997.2888 364.6434 1997.2916 364.6414 1997.2943 364.6393 1997.2971 364.6373 1997.2998 364.6352 1997.3025 364.6332 1997.3053 364.6312 1997.308 364.6293 1997.3107 364.6273 1997.3135 364.6254 1997.3162 364.6235 1997.319 364.6217 1997.3217 364.6198 1997.3244 364.618 1997.3272 364.6162 1997.3299 364.6144 1997.3326 364.6126 1997.3354 364.6109 1997.3381 364.6091 1997.3409 364.6075 1997.3436 364.6058 1997.3463 364.6041 1997.3491 364.6025 1997.3518 364.6009 1997.3546 364.5994 1997.3573 364.5978 1997.36 364.5963 1997.3628 364.5948 1997.3655 364.5933 1997.3682 364.5919 1997.371 364.5905 1997.3737 364.5891 1997.3765 364.5878 1997.3792 364.5865 1997.3819 364.5852 1997.3847 364.5839 1997.3874 364.5826 1997.3901 364.5814 1997.3929 364.5803 1997.3956 364.5791 1997.3984 364.578 1997.4011 364.5769 1997.4038 364.5758 1997.4066 364.5748 1997.4093 364.5738 1997.412 364.5728 1997.4148 364.5719 1997.4175 364.571 1997.4203 364.5702 1997.423 364.5693 1997.4257 364.5685 1997.4285 364.5678 1997.4312 364.567 1997.4339 364.5663 1997.4367 364.5657 1997.4394 364.565 1997.4422 364.5644 1997.4449 364.5638 1997.4476 364.5633 1997.4504 364.5629 1997.4531 364.5625 1997.4559 364.5621 1997.4586 364.5617 1997.4613 364.5613 1997.4641 364.561 1997.4668 364.5607 1997.4695 364.5605 1997.4723 364.5603 1997.475 364.5601 1997.4778 364.56 1997.4805 364.5599 1997.4832 364.5598 1997.486 364.5598 1997.4887 364.56 1997.4914 364.5601 1997.4942 364.5602 1997.4969 364.5604 1997.4997 364.5606 1997.5024 364.5608 1997.5051 364.5611 1997.5079 364.5614 1997.5106 364.5617 1997.5133 364.5621 1997.5161 364.5626 1997.5188 364.563 1997.5216 364.5635 1997.5243 364.5641 1997.527 364.5646 1997.5298 364.5656 1997.5325 364.5663 1997.5352 364.567 1997.538 364.5678 1997.5407 364.5686 1997.5435 364.5695 1997.5462 364.5704 1997.5489 364.5714 1997.5517 364.5724 1997.5544 364.5734 1997.5572 364.5745 1997.5599 364.5756 1997.5626 364.5767 1997.5654 364.578 1997.5681 364.5795 1997.5708 364.5809 1997.5736 364.5822 1997.5763 364.5837 1997.5791 364.5851 1997.5818 364.5866 1997.5845 364.5882 1997.5873 364.5898 1997.59 364.5914 1997.5927 364.5931 1997.5955 364.5948 1997.5982 364.5966 1997.601 364.5984 1997.6037 364.6002 1997.6064 364.6025 1997.6092 364.6045 1997.6119 364.6065 1997.6146 364.6086 1997.6174 364.6108 1997.6201 364.6129 1997.6229 364.6152 1997.6256 364.6174 1997.6283 364.6198 1997.6311 364.6221 1997.6338 364.6245 1997.6366 364.627 1997.6393 364.6295 1997.642 364.632 1997.6448 364.6346 1997.6475 364.6372 1997.6502 364.6399 1997.653 364.6433 1997.6557 364.6461 1997.6585 364.649 1997.6612 364.652 1997.6639 364.6549 1997.6667 364.658 1997.6694 364.6611 1997.6721 364.6642 1997.6749 364.6674 1997.6776 364.6706 1997.6804 364.6739 1997.6831 364.6772 1997.6858 364.6805 1997.6886 364.684 1997.6913 364.6874 1997.694 364.6909 1997.6968 364.6945 1997.6995 364.6988 1997.7023 364.7025 1997.705 364.7063 1997.7077 364.7101 1997.7105 364.714 1997.7132 364.7179 1997.7159 364.7219 1997.7187 364.7259 1997.7214 364.73 1997.7242 364.7341 1997.7269 364.7383 1997.7296 364.7425 1997.7324 364.7468 1997.7351 364.7511 1997.7379 364.7559 1997.7406 364.7604 1997.7433 364.765 1997.7461 364.7695 1997.7488 364.7742 1997.7515 364.7788 1997.7543 364.7836 1997.757 364.7883 1997.7598 364.7932 1997.7625 364.798 1997.7652 364.8029 1997.768 364.8079 1997.7707 364.8129 1997.7734 364.818 1997.7762 364.8231 1997.7789 364.8289 1997.7817 364.8342 1997.7844 364.8395 1997.7871 364.8449 1997.7899 364.8504 1997.7926 364.8559 1997.7953 364.8614 1997.7981 364.867 1997.8008 364.8726 1997.8036 364.8783 1997.8063 364.8841 1997.809 364.8898 1997.8118 364.8957 1997.8145 364.9016 1997.8172 364.9081 1997.82 364.9141 1997.8227 364.9202 1997.8255 364.9264 1997.8282 364.9326 1997.8309 364.9389 1997.8337 364.9452 1997.8364 364.9515 1997.8392 364.9579 1997.8419 364.9644 1997.8446 364.9709 1997.8474 364.9774 1997.8501 364.984 1997.8528 364.9911 1997.8556 364.9979 1997.8583 365.0047 1997.8611 365.0116 1997.8638 365.0185 1997.8665 365.0255 1997.8693 365.0325 1997.872 365.0395 1997.8747 365.0466 1997.8775 365.0538 1997.8802 365.061 1997.883 365.0682 1997.8857 365.0755 1997.8884 365.0829 1997.8912 365.0909 1997.8939 365.0984 1997.8966 365.106 1997.8994 365.1136 1997.9021 365.1212 1997.9049 365.129 1997.9076 365.1367 1997.9103 365.1445 1997.9131 365.1524 1997.9158 365.1603 1997.9185 365.1682 1997.9213 365.1762 1997.924 365.1843 1997.9268 365.1929 1997.9295 365.2011 1997.9322 365.2093 1997.935 365.2176 1997.9377 365.226 1997.9405 365.2344 1997.9432 365.2428 1997.9459 365.2513 1997.9487 365.2598 1997.9514 365.2684 1997.9541 365.277 1997.9569 365.2857 1997.9596 365.2944 1997.9624 365.3032 1997.9651 365.312 1997.9678 365.3208 1997.9706 365.3305 1997.9733 365.3395 1997.976 365.3486 1997.9788 365.3577 1997.9815 365.3669 1997.9843 365.3761 1997.987 365.3853 1997.9897 365.3946 1997.9925 365.404 1997.9952 365.4133 1997.9979 365.4228 1998.0007 365.4322 1998.0034 365.4417 1998.0062 365.4513 1998.0089 365.4609 1998.0116 365.4705 1998.0144 365.4802 1998.0171 365.4899 1998.0198 365.4997 1998.0226 365.5095 1998.0253 365.5194 1998.0281 365.5292 1998.0308 365.5392 1998.0335 365.5491 1998.0363 365.5592 1998.039 365.5692 1998.0418 365.5793 1998.0445 365.5894 1998.0472 365.6018 1998.05 365.6122 1998.0527 365.6226 1998.0554 365.6331 1998.0582 365.6436 1998.0609 365.6541 1998.0637 365.6647 1998.0664 365.6753 1998.0691 365.6859 1998.0719 365.6966 1998.0746 365.7073 1998.0773 365.7181 1998.0801 365.7289 1998.0828 365.7397 1998.0856 365.7511 1998.0883 365.7621 1998.091 365.7731 1998.0938 365.7842 1998.0965 365.7953 1998.0992 365.8064 1998.102 365.8176 1998.1047 365.8288 1998.1075 365.84 1998.1102 365.8513 1998.1129 365.8626 1998.1157 365.8739 1998.1184 365.8853 1998.1211 365.8967 1998.1239 365.9087 1998.1266 365.9202 1998.1294 365.9318 1998.1321 365.9434 1998.1348 365.9551 1998.1376 365.9668 1998.1403 365.9785 1998.1431 365.9902 1998.1458 366.002 1998.1485 366.0139 1998.1513 366.0257 1998.154 366.0376 1998.1567 366.0495 1998.1595 366.0615 1998.1622 366.0739 1998.165 366.0859 1998.1677 366.098 1998.1704 366.1102 1998.1732 366.1223 1998.1759 366.1345 1998.1786 366.1468 1998.1814 366.159 1998.1841 366.1713 1998.1869 366.1836 1998.1896 366.196 1998.1923 366.2083 1998.1951 366.2207 1998.1978 366.2332 1998.2005 366.246 1998.2033 366.2585 1998.206 366.2711 1998.2088 366.2837 1998.2115 366.2963 1998.2142 366.3089 1998.217 366.3216 1998.2197 366.3343 1998.2225 366.347 1998.2252 366.3598 1998.2279 366.3726 1998.2307 366.3854 1998.2334 366.3982 1998.2361 366.411 1998.2389 366.4239 1998.2416 366.4368 1998.2444 366.4501 1998.2471 366.4631 1998.2498 366.4761 1998.2526 366.4892 1998.2553 366.5022 1998.258 366.5153 1998.2608 366.5284 1998.2635 366.5415 1998.2663 366.5547 1998.269 366.5679 1998.2717 366.5811 1998.2745 366.5943 1998.2772 366.6075 1998.2799 366.6208 1998.2827 366.6341 1998.2854 366.6476 1998.2882 366.6609 1998.2909 366.6743 1998.2936 366.6877 1998.2964 366.7011 1998.2991 366.7145 1998.3018 366.728 1998.3046 366.7414 1998.3073 366.7549 1998.3101 366.7684 1998.3128 366.7819 1998.3155 366.7954 1998.3183 366.809 1998.321 366.8226 1998.3238 366.8361 1998.3265 366.8499 1998.3292 366.8635 1998.332 366.8771 1998.3347 366.8908 1998.3374 366.9045 1998.3402 366.9182 1998.3429 366.9319 1998.3457 366.9456 1998.3484 366.9593 1998.3511 366.9731 1998.3539 366.9868 1998.3566 367.0006 1998.3593 367.0144 1998.3621 367.0282 1998.3648 367.042 1998.3676 367.0559 1998.3703 367.0697 1998.373 367.0836 1998.3758 367.0974 1998.3785 367.1113 1998.3812 367.1252 1998.384 367.1391 1998.3867 367.153 1998.3895 367.1669 1998.3922 367.1808 1998.3949 367.1947 1998.3977 367.2087 1998.4004 367.2226 1998.4031 367.2366 1998.4059 367.2505 1998.4086 367.2645 1998.4114 367.2785 1998.4141 367.2924 1998.4168 367.3064 1998.4196 367.3204 1998.4223 367.3344 1998.4251 367.3484 1998.4278 367.3624 1998.4305 367.3764 1998.4333 367.3904 1998.436 367.4045 1998.4387 367.4185 1998.4415 367.4325 1998.4442 367.4465 1998.447 367.4606 1998.4497 367.4746 1998.4524 367.4887 1998.4552 367.5027 1998.4579 367.5168 1998.4606 367.5308 1998.4634 367.5449 1998.4661 367.5589 1998.4689 367.5729 1998.4716 367.5869 1998.4743 367.6009 1998.4771 367.615 1998.4798 367.629 1998.4825 367.6431 1998.4853 367.6571 1998.488 367.6711 1998.4908 367.6852 1998.4935 367.6992 1998.4962 367.7132 1998.499 367.7273 1998.5017 367.7413 1998.5044 367.7553 1998.5072 367.7692 1998.5099 367.7832 1998.5127 367.7972 1998.5154 367.8112 1998.5181 367.8252 1998.5209 367.8392 1998.5236 367.8531 1998.5264 367.8671 1998.5291 367.8811 1998.5318 367.8951 1998.5346 367.909 1998.5373 367.923 1998.54 367.9369 1998.5428 367.9509 1998.5455 367.9646 1998.5483 367.9785 1998.551 367.9924 1998.5537 368.0063 1998.5565 368.0202 1998.5592 368.034 1998.5619 368.0479 1998.5647 368.0617 1998.5674 368.0756 1998.5702 368.0894 1998.5729 368.1032 1998.5756 368.1171 1998.5784 368.1309 1998.5811 368.1447 1998.5838 368.1582 1998.5866 368.172 1998.5893 368.1857 1998.5921 368.1994 1998.5948 368.2131 1998.5975 368.2268 1998.6003 368.2405 1998.603 368.2542 1998.6057 368.2678 1998.6085 368.2814 1998.6112 368.2951 1998.614 368.3087 1998.6167 368.3223 1998.6194 368.3359 1998.6222 368.3495 1998.6249 368.3627 1998.6277 368.3762 1998.6304 368.3897 1998.6331 368.4032 1998.6359 368.4166 1998.6386 368.4301 1998.6413 368.4435 1998.6441 368.4569 1998.6468 368.4703 1998.6496 368.4837 1998.6523 368.497 1998.655 368.5104 1998.6578 368.5237 1998.6605 368.537 1998.6632 368.5503 1998.666 368.5636 1998.6687 368.5768 1998.6715 368.5896 1998.6742 368.6028 1998.6769 368.6159 1998.6797 368.629 1998.6824 368.6421 1998.6851 368.6552 1998.6879 368.6683 1998.6906 368.6813 1998.6934 368.6943 1998.6961 368.7073 1998.6988 368.7203 1998.7016 368.7332 1998.7043 368.7462 1998.707 368.7591 1998.7098 368.7716 1998.7125 368.7844 1998.7153 368.7972 1998.718 368.81 1998.7207 368.8227 1998.7235 368.8354 1998.7262 368.8481 1998.729 368.8608 1998.7317 368.8734 1998.7344 368.8861 1998.7372 368.8987 1998.7399 368.9112 1998.7426 368.9238 1998.7454 368.9363 1998.7481 368.9488 1998.7509 368.9613 1998.7536 368.9737 1998.7563 368.9862 1998.7591 368.9986 1998.7618 369.0109 1998.7645 369.0233 1998.7673 369.0356 1998.77 369.0479 1998.7728 369.0601 1998.7755 369.0724 1998.7782 369.0846 1998.781 369.0968 1998.7837 369.1089 1998.7864 369.121 1998.7892 369.1331 1998.7919 369.1452 1998.7947 369.1572 1998.7974 369.1692 1998.8001 369.1812 1998.8029 369.1932 1998.8056 369.2051 1998.8084 369.217 1998.8111 369.2288 1998.8138 369.2407 1998.8166 369.2525 1998.8193 369.2642 1998.822 369.276 1998.8248 369.2877 1998.8275 369.2993 1998.8303 369.311 1998.833 369.3226 1998.8357 369.3342 1998.8385 369.3457 1998.8412 369.3572 1998.8439 369.3637 1998.8467 369.3749 1998.8494 369.3862 1998.8522 369.3973 1998.8549 369.4085 1998.8576 369.4196 1998.8604 369.4307 1998.8631 369.4417 1998.8658 369.4527 1998.8686 369.4637 1998.8713 369.4746 1998.8741 369.4855 1998.8768 369.4964 1998.8795 369.5073 1998.8823 369.5181 1998.885 369.5283 1998.8877 369.539 1998.8905 369.5496 1998.8932 369.5602 1998.896 369.5708 1998.8987 369.5813 1998.9014 369.5918 1998.9042 369.6022 1998.9069 369.6127 1998.9097 369.623 1998.9124 369.6334 1998.9151 369.6437 1998.9179 369.654 1998.9206 369.6642 1998.9233 369.6744 1998.9261 369.684 1998.9288 369.6941 1998.9316 369.7041 1998.9343 369.7141 1998.937 369.724 1998.9398 369.7339 1998.9425 369.7438 1998.9452 369.7537 1998.948 369.7635 1998.9507 369.7732 1998.9535 369.7829 1998.9562 369.7926 1998.9589 369.8023 1998.9617 369.8119 1998.9644 369.8214 1998.9671 369.8304 1998.9699 369.8398 1998.9726 369.8492 1998.9754 369.8586 1998.9781 369.8679 1998.9808 369.8771 1998.9836 369.8863 1998.9863 369.8955 1998.989 369.9047 1998.9918 369.9138 1998.9945 369.9229 1998.9973 369.9319 1999 369.9409 1999.0027 369.9498 1999.0055 369.9583 1999.0082 369.9671 1999.011 369.9758 1999.0137 369.9846 1999.0164 369.9932 1999.0192 370.0019 1999.0219 370.0105 1999.0246 370.0191 1999.0274 370.0276 1999.0301 370.0361 1999.0329 370.0445 1999.0356 370.0529 1999.0383 370.0613 1999.0411 370.0696 1999.0438 370.0774 1999.0465 370.0856 1999.0493 370.0937 1999.052 370.1018 1999.0548 370.1098 1999.0575 370.1179 1999.0602 370.1258 1999.063 370.1338 1999.0657 370.1417 1999.0684 370.1495 1999.0712 370.1573 1999.0739 370.1651 1999.0767 370.1728 1999.0794 370.1805 1999.0821 370.1876 1999.0849 370.1952 1999.0876 370.2026 1999.0903 370.2101 1999.0931 370.2175 1999.0958 370.2249 1999.0986 370.2322 1999.1013 370.2395 1999.104 370.2467 1999.1068 370.2539 1999.1095 370.2611 1999.1123 370.2682 1999.115 370.2753 1999.1177 370.2823 1999.1205 370.2893 1999.1232 370.2963 1999.1259 370.3032 1999.1287 370.3101 1999.1314 370.3169 1999.1342 370.3237 1999.1369 370.3304 1999.1396 370.336 1999.1424 370.3425 1999.1451 370.349 1999.1478 370.3555 1999.1506 370.3619 1999.1533 370.3683 1999.1561 370.3747 1999.1588 370.381 1999.1615 370.3873 1999.1643 370.3933 1999.167 370.3994 1999.1697 370.4055 1999.1725 370.4116 1999.1752 370.4176 1999.178 370.4235 1999.1807 370.4295 1999.1834 370.4354 1999.1862 370.4412 1999.1889 370.447 1999.1916 370.4528 1999.1944 370.4585 1999.1971 370.4642 1999.1999 370.4699 1999.2026 370.4755 1999.2053 370.481 1999.2081 370.4858 1999.2108 370.4912 1999.2136 370.4966 1999.2163 370.5019 1999.219 370.5072 1999.2218 370.5124 1999.2245 370.5176 1999.2272 370.5227 1999.23 370.5279 1999.2327 370.5329 1999.2355 370.538 1999.2382 370.543 1999.2409 370.5479 1999.2437 370.5523 1999.2464 370.5571 1999.2491 370.5619 1999.2519 370.5666 1999.2546 370.5713 1999.2574 370.5759 1999.2601 370.5805 1999.2628 370.5851 1999.2656 370.5896 1999.2683 370.5941 1999.271 370.5985 1999.2738 370.6029 1999.2765 370.6073 1999.2793 370.6116 1999.282 370.6153 1999.2847 370.6195 1999.2875 370.6236 1999.2902 370.6277 1999.293 370.6318 1999.2957 370.6358 1999.2984 370.6398 1999.3012 370.6437 1999.3039 370.6476 1999.3066 370.6514 1999.3094 370.6553 1999.3121 370.659 1999.3149 370.6628 1999.3176 370.6665 1999.3203 370.6701 1999.3231 370.6738 1999.3258 370.6774 1999.3285 370.6809 1999.3313 370.6835 1999.334 370.6868 1999.3368 370.6902 1999.3395 370.6935 1999.3422 370.6967 1999.345 370.6999 1999.3477 370.7031 1999.3504 370.7063 1999.3532 370.7094 1999.3559 370.7125 1999.3587 370.7155 1999.3614 370.7185 1999.3641 370.7215 1999.3669 370.7244 1999.3696 370.7273 1999.3723 370.7294 1999.3751 370.7322 1999.3778 370.7349 1999.3806 370.7375 1999.3833 370.7401 1999.386 370.7427 1999.3888 370.7453 1999.3915 370.7478 1999.3943 370.7503 1999.397 370.7527 1999.3997 370.7551 1999.4025 370.7575 1999.4052 370.7599 1999.4079 370.7616 1999.4107 370.7638 1999.4134 370.766 1999.4162 370.7681 1999.4189 370.7702 1999.4216 370.7722 1999.4244 370.7742 1999.4271 370.7762 1999.4298 370.7782 1999.4326 370.7801 1999.4353 370.782 1999.4381 370.7839 1999.4408 370.7857 1999.4435 370.7875 1999.4463 370.7886 1999.449 370.7902 1999.4517 370.7918 1999.4545 370.7934 1999.4572 370.7949 1999.46 370.7964 1999.4627 370.7979 1999.4654 370.7993 1999.4682 370.8007 1999.4709 370.8021 1999.4736 370.8035 1999.4764 370.8048 1999.4791 370.8061 1999.4819 370.8073 1999.4846 370.8078 1999.4873 370.8089 1999.4901 370.81 1999.4928 370.8111 1999.4956 370.8121 1999.4983 370.813 1999.501 370.814 1999.5038 370.8149 1999.5065 370.8158 1999.5092 370.8167 1999.512 370.8175 1999.5147 370.8183 1999.5175 370.8191 1999.5202 370.8198 1999.5229 370.8198 1999.5257 370.8204 1999.5284 370.821 1999.5311 370.8215 1999.5339 370.822 1999.5366 370.8225 1999.5394 370.823 1999.5421 370.8234 1999.5448 370.8238 1999.5476 370.8242 1999.5503 370.8245 1999.553 370.8248 1999.5558 370.8251 1999.5585 370.8254 1999.5613 370.8249 1999.564 370.825 1999.5667 370.8251 1999.5695 370.8251 1999.5722 370.8252 1999.5749 370.8252 1999.5777 370.8252 1999.5804 370.8251 1999.5832 370.8251 1999.5859 370.825 1999.5886 370.8249 1999.5914 370.8248 1999.5941 370.8246 1999.5969 370.8244 1999.5996 370.8234 1999.6023 370.8231 1999.6051 370.8227 1999.6078 370.8223 1999.6105 370.8219 1999.6133 370.8215 1999.616 370.8211 1999.6188 370.8206 1999.6215 370.8201 1999.6242 370.8196 1999.627 370.8191 1999.6297 370.8185 1999.6324 370.8179 1999.6352 370.8173 1999.6379 370.8159 1999.6407 370.8151 1999.6434 370.8144 1999.6461 370.8136 1999.6489 370.8128 1999.6516 370.8119 1999.6543 370.8111 1999.6571 370.8102 1999.6598 370.8093 1999.6626 370.8084 1999.6653 370.8075 1999.668 370.8066 1999.6708 370.8056 1999.6735 370.8047 1999.6762 370.8028 1999.679 370.8017 1999.6817 370.8005 1999.6845 370.7994 1999.6872 370.7982 1999.6899 370.797 1999.6927 370.7958 1999.6954 370.7946 1999.6982 370.7933 1999.7009 370.7921 1999.7036 370.7908 1999.7064 370.7895 1999.7091 370.7882 1999.7118 370.7869 1999.7146 370.7847 1999.7173 370.7832 1999.7201 370.7818 1999.7228 370.7803 1999.7255 370.7787 1999.7283 370.7772 1999.731 370.7757 1999.7337 370.7741 1999.7365 370.7726 1999.7392 370.771 1999.742 370.7694 1999.7447 370.7678 1999.7474 370.7662 1999.7502 370.7646 1999.7529 370.7621 1999.7556 370.7603 1999.7584 370.7586 1999.7611 370.7568 1999.7639 370.755 1999.7666 370.7532 1999.7693 370.7513 1999.7721 370.7495 1999.7748 370.7477 1999.7775 370.7458 1999.7803 370.7439 1999.783 370.7421 1999.7858 370.7402 1999.7885 370.7383 1999.7912 370.7364 1999.794 370.7345 1999.7967 370.7326 1999.7995 370.7307 1999.8022 370.7287 1999.8049 370.7268 1999.8077 370.7249 1999.8104 370.7229 1999.8131 370.7209 1999.8159 370.719 1999.8186 370.717 1999.8214 370.715 1999.8241 370.713 1999.8268 370.711 1999.8296 370.709 1999.8323 370.707 1999.835 370.705 1999.8378 370.703 1999.8405 370.701 1999.8433 370.699 1999.846 370.6969 1999.8487 370.6949 1999.8515 370.6929 1999.8542 370.6908 1999.8569 370.6888 1999.8597 370.6867 1999.8624 370.6847 1999.8652 370.6826 1999.8679 370.6805 1999.8706 370.6785 1999.8734 370.6764 1999.8761 370.6743 1999.8789 370.6723 1999.8816 370.6702 1999.8843 370.6681 1999.8871 370.666 1999.8898 370.6639 1999.8925 370.6619 1999.8953 370.6598 1999.898 370.6577 1999.9008 370.6556 1999.9035 370.6535 1999.9062 370.6389 1999.909 370.6363 1999.9117 370.6338 1999.9144 370.6313 1999.9172 370.6288 1999.9199 370.6262 1999.9227 370.6237 1999.9254 370.6212 1999.9281 370.6187 1999.9309 370.6161 1999.9336 370.6136 1999.9363 370.6111 1999.9391 370.6086 1999.9418 370.6061 1999.9446 370.6029 1999.9473 370.6003 1999.95 370.5977 1999.9528 370.595 1999.9555 370.5924 1999.9582 370.5899 1999.961 370.5873 1999.9637 370.5847 1999.9665 370.5821 1999.9692 370.5795 1999.9719 370.5769 1999.9747 370.5743 1999.9774 370.5718 1999.9802 370.5692 1999.9829 370.566 1999.9856 370.5633 1999.9884 370.5607 1999.9911 370.558 1999.9938 370.5554 1999.9966 370.5528 1999.9993 370.5502 2000.0021 370.5475 2000.0048 370.5449 2000.0075 370.5423 2000.0103 370.5397 2000.013 370.5371 2000.0157 370.5345 2000.0185 370.532 2000.0212 370.5288 2000.024 370.5261 2000.0267 370.5235 2000.0294 370.5209 2000.0322 370.5182 2000.0349 370.5156 2000.0376 370.513 2000.0404 370.5104 2000.0431 370.5078 2000.0459 370.5052 2000.0486 370.5026 2000.0513 370.5001 2000.0541 370.4975 2000.0568 370.4949 2000.0595 370.4919 2000.0623 370.4892 2000.065 370.4866 2000.0678 370.4841 2000.0705 370.4815 2000.0732 370.4789 2000.076 370.4763 2000.0787 370.4738 2000.0815 370.4713 2000.0842 370.4687 2000.0869 370.4662 2000.0897 370.4637 2000.0924 370.4612 2000.0951 370.4587 2000.0979 370.4563 2000.1006 370.4538 2000.1034 370.4513 2000.1061 370.4489 2000.1088 370.4465 2000.1116 370.4441 2000.1143 370.4417 2000.117 370.4382 2000.1198 370.4358 2000.1225 370.4333 2000.1253 370.4309 2000.128 370.4284 2000.1307 370.426 2000.1335 370.4236 2000.1362 370.4212 2000.1389 370.4188 2000.1417 370.4165 2000.1444 370.4141 2000.1472 370.4118 2000.1499 370.4095 2000.1526 370.4071 2000.1554 370.4048 2000.1581 370.4021 2000.1608 370.3998 2000.1636 370.3975 2000.1663 370.3952 2000.1691 370.393 2000.1718 370.3907 2000.1745 370.3884 2000.1773 370.3862 2000.18 370.384 2000.1828 370.3818 2000.1855 370.3796 2000.1882 370.3775 2000.191 370.3753 2000.1937 370.3729 2000.1964 370.3707 2000.1992 370.3686 2000.2019 370.3665 2000.2047 370.3644 2000.2074 370.3623 2000.2101 370.3602 2000.2129 370.3581 2000.2156 370.3561 2000.2183 370.3541 2000.2211 370.3521 2000.2238 370.3501 2000.2266 370.3481 2000.2293 370.3462 2000.232 370.3442 2000.2348 370.342 2000.2375 370.3401 2000.2402 370.3382 2000.243 370.3363 2000.2457 370.3344 2000.2485 370.3325 2000.2512 370.3307 2000.2539 370.3288 2000.2567 370.327 2000.2594 370.3252 2000.2621 370.3235 2000.2649 370.3217 2000.2676 370.32 2000.2704 370.3181 2000.2731 370.3164 2000.2758 370.3147 2000.2786 370.313 2000.2813 370.3113 2000.2841 370.3097 2000.2868 370.3081 2000.2895 370.3065 2000.2923 370.3049 2000.295 370.3033 2000.2977 370.3018 2000.3005 370.3002 2000.3032 370.2987 2000.306 370.2973 2000.3087 370.2956 2000.3114 370.2942 2000.3142 370.2927 2000.3169 370.2913 2000.3196 370.2899 2000.3224 370.2885 2000.3251 370.2872 2000.3279 370.2858 2000.3306 370.2845 2000.3333 370.2832 2000.3361 370.2819 2000.3388 370.2807 2000.3415 370.2795 2000.3443 370.2782 2000.347 370.2769 2000.3498 370.2758 2000.3525 370.2746 2000.3552 370.2734 2000.358 370.2723 2000.3607 370.2712 2000.3634 370.2701 2000.3662 370.2691 2000.3689 370.2681 2000.3717 370.267 2000.3744 370.2661 2000.3771 370.2651 2000.3799 370.2641 2000.3826 370.2632 2000.3854 370.2623 2000.3881 370.2614 2000.3908 370.2606 2000.3936 370.2597 2000.3963 370.2589 2000.399 370.2581 2000.4018 370.2572 2000.4045 370.2565 2000.4073 370.2557 2000.41 370.255 2000.4127 370.2543 2000.4155 370.2537 2000.4182 370.253 2000.4209 370.2524 2000.4237 370.2518 2000.4264 370.2512 2000.4292 370.2507 2000.4319 370.2501 2000.4346 370.2496 2000.4374 370.2491 2000.4401 370.2487 2000.4428 370.2482 2000.4456 370.2478 2000.4483 370.2474 2000.4511 370.247 2000.4538 370.2467 2000.4565 370.2464 2000.4593 370.2461 2000.462 370.2458 2000.4648 370.2455 2000.4675 370.2453 2000.4702 370.2451 2000.473 370.2449 2000.4757 370.2447 2000.4784 370.2446 2000.4812 370.2445 2000.4839 370.2444 2000.4867 370.2443 2000.4894 370.2443 2000.4921 370.2442 2000.4949 370.2442 2000.4976 370.2443 2000.5003 370.2443 2000.5031 370.2444 2000.5058 370.2445 2000.5086 370.2446 2000.5113 370.2448 2000.514 370.2449 2000.5168 370.2451 2000.5195 370.2453 2000.5222 370.2456 2000.525 370.2458 2000.5277 370.2461 2000.5305 370.2464 2000.5332 370.2468 2000.5359 370.2472 2000.5387 370.2476 2000.5414 370.248 2000.5441 370.2484 2000.5469 370.2489 2000.5496 370.2493 2000.5524 370.2498 2000.5551 370.2504 2000.5578 370.2509 2000.5606 370.2515 2000.5633 370.2521 2000.5661 370.2527 2000.5688 370.2534 2000.5715 370.254 2000.5743 370.2547 2000.577 370.2555 2000.5797 370.2563 2000.5825 370.257 2000.5852 370.2578 2000.588 370.2587 2000.5907 370.2595 2000.5934 370.2604 2000.5962 370.2613 2000.5989 370.2622 2000.6016 370.2631 2000.6044 370.2641 2000.6071 370.2651 2000.6099 370.2661 2000.6126 370.2671 2000.6153 370.2681 2000.6181 370.2692 2000.6208 370.2703 2000.6235 370.2714 2000.6263 370.2726 2000.629 370.2738 2000.6318 370.275 2000.6345 370.2762 2000.6372 370.2774 2000.64 370.2787 2000.6427 370.28 2000.6454 370.2813 2000.6482 370.2826 2000.6509 370.284 2000.6537 370.2853 2000.6564 370.2867 2000.6591 370.2882 2000.6619 370.2896 2000.6646 370.2911 2000.6674 370.2926 2000.6701 370.2941 2000.6728 370.2962 2000.6756 370.2978 2000.6783 370.2994 2000.681 370.301 2000.6838 370.3027 2000.6865 370.3044 2000.6893 370.3061 2000.692 370.3078 2000.6947 370.3096 2000.6975 370.3113 2000.7002 370.3131 2000.7029 370.3149 2000.7057 370.3168 2000.7084 370.3186 2000.7112 370.3207 2000.7139 370.3226 2000.7166 370.3245 2000.7194 370.3265 2000.7221 370.3285 2000.7248 370.3305 2000.7276 370.3325 2000.7303 370.3346 2000.7331 370.3367 2000.7358 370.3388 2000.7385 370.3409 2000.7413 370.343 2000.744 370.3452 2000.7467 370.3475 2000.7495 370.3497 2000.7522 370.3519 2000.755 370.3542 2000.7577 370.3565 2000.7604 370.3588 2000.7632 370.3611 2000.7659 370.3635 2000.7687 370.3659 2000.7714 370.3683 2000.7741 370.3707 2000.7769 370.3731 2000.7796 370.3756 2000.7823 370.378 2000.7851 370.3805 2000.7878 370.3832 2000.7906 370.3858 2000.7933 370.3884 2000.796 370.391 2000.7988 370.3936 2000.8015 370.3962 2000.8042 370.3989 2000.807 370.4016 2000.8097 370.4043 2000.8125 370.407 2000.8152 370.4097 2000.8179 370.4125 2000.8207 370.4153 2000.8234 370.4181 2000.8261 370.4211 2000.8289 370.4239 2000.8316 370.4268 2000.8344 370.4297 2000.8371 370.4326 2000.8398 370.4356 2000.8426 370.4385 2000.8453 370.4415 2000.848 370.4445 2000.8508 370.4475 2000.8535 370.4505 2000.8563 370.4536 2000.859 370.4567 2000.8617 370.4597 2000.8645 370.4628 2000.8672 370.466 2000.87 370.4691 2000.8727 370.4723 2000.8754 370.4755 2000.8782 370.4787 2000.8809 370.4819 2000.8836 370.4851 2000.8864 370.4883 2000.8891 370.4916 2000.8919 370.4949 2000.8946 370.4988 2000.8973 370.5022 2000.9001 370.5055 2000.9028 370.5089 2000.9055 370.5124 2000.9083 370.5158 2000.911 370.5192 2000.9138 370.5227 2000.9165 370.5262 2000.9192 370.5297 2000.922 370.5332 2000.9247 370.5368 2000.9274 370.5403 2000.9302 370.5439 2000.9329 370.5475 2000.9357 370.551 2000.9384 370.5547 2000.9411 370.5586 2000.9439 370.5623 2000.9466 370.566 2000.9493 370.5697 2000.9521 370.5734 2000.9548 370.5771 2000.9576 370.5809 2000.9603 370.5847 2000.963 370.5885 2000.9658 370.5923 2000.9685 370.5961 2000.9713 370.5999 2000.974 370.6038 2000.9767 370.6076 2000.9795 370.6115 2000.9822 370.6154 2000.9849 370.6193 2000.9877 370.6232 2000.9904 370.6272 2000.9932 370.6311 2000.9959 370.6351 2000.9986 370.639 2001.0014 370.643 2001.0041 370.647 2001.0068 370.6511 2001.0096 370.6551 2001.0123 370.6591 2001.0151 370.6632 2001.0178 370.6673 2001.0205 370.6713 2001.0233 370.6754 2001.026 370.6796 2001.0287 370.6837 2001.0315 370.6878 2001.0342 370.692 2001.037 370.6961 2001.0397 370.7003 2001.0424 370.7045 2001.0452 370.7087 2001.0479 370.7129 2001.0507 370.7171 2001.0534 370.7213 2001.0561 370.7276 2001.0589 370.7319 2001.0616 370.7363 2001.0643 370.7407 2001.0671 370.7451 2001.0698 370.7495 2001.0726 370.7539 2001.0753 370.7583 2001.078 370.7627 2001.0808 370.7672 2001.0835 370.7716 2001.0862 370.7761 2001.089 370.7806 2001.0917 370.785 2001.0945 370.7898 2001.0972 370.7943 2001.0999 370.7989 2001.1027 370.8035 2001.1054 370.808 2001.1081 370.8126 2001.1109 370.8172 2001.1136 370.8218 2001.1164 370.8264 2001.1191 370.831 2001.1218 370.8357 2001.1246 370.8403 2001.1273 370.845 2001.13 370.8496 2001.1328 370.8546 2001.1355 370.8593 2001.1383 370.864 2001.141 370.8687 2001.1437 370.8735 2001.1465 370.8782 2001.1492 370.883 2001.152 370.8877 2001.1547 370.8925 2001.1574 370.8973 2001.1602 370.9021 2001.1629 370.9069 2001.1656 370.9117 2001.1684 370.9165 2001.1711 370.9216 2001.1739 370.9265 2001.1766 370.9314 2001.1793 370.9362 2001.1821 370.9411 2001.1848 370.946 2001.1875 370.9509 2001.1903 370.9559 2001.193 370.9608 2001.1958 370.9657 2001.1985 370.9706 2001.2012 370.9756 2001.204 370.9805 2001.2067 370.9855 2001.2094 370.9907 2001.2122 370.9957 2001.2149 371.0008 2001.2177 371.0058 2001.2204 371.0108 2001.2231 371.0158 2001.2259 371.0209 2001.2286 371.0259 2001.2313 371.0309 2001.2341 371.036 2001.2368 371.041 2001.2396 371.0461 2001.2423 371.0512 2001.245 371.0562 2001.2478 371.0616 2001.2505 371.0668 2001.2533 371.0719 2001.256 371.077 2001.2587 371.0822 2001.2615 371.0873 2001.2642 371.0925 2001.2669 371.0976 2001.2697 371.1028 2001.2724 371.1079 2001.2752 371.1131 2001.2779 371.1182 2001.2806 371.1234 2001.2834 371.1286 2001.2861 371.1341 2001.2888 371.1393 2001.2916 371.1445 2001.2943 371.1498 2001.2971 371.155 2001.2998 371.1602 2001.3025 371.1655 2001.3053 371.1707 2001.308 371.176 2001.3107 371.1812 2001.3135 371.1865 2001.3162 371.1917 2001.319 371.197 2001.3217 371.2022 2001.3244 371.2078 2001.3272 371.2131 2001.3299 371.2184 2001.3326 371.2237 2001.3354 371.229 2001.3381 371.2343 2001.3409 371.2397 2001.3436 371.245 2001.3463 371.2503 2001.3491 371.2556 2001.3518 371.2609 2001.3546 371.2662 2001.3573 371.2716 2001.36 371.2769 2001.3628 371.2822 2001.3655 371.2875 2001.3682 371.2928 2001.371 371.2987 2001.3737 371.304 2001.3765 371.3094 2001.3792 371.3148 2001.3819 371.3202 2001.3847 371.3256 2001.3874 371.3309 2001.3901 371.3363 2001.3929 371.3417 2001.3956 371.3471 2001.3984 371.3525 2001.4011 371.3578 2001.4038 371.3632 2001.4066 371.3686 2001.4093 371.374 2001.412 371.3793 2001.4148 371.3847 2001.4175 371.3901 2001.4203 371.3955 2001.423 371.4008 2001.4257 371.4062 2001.4285 371.4123 2001.4312 371.4177 2001.4339 371.4232 2001.4367 371.4286 2001.4394 371.434 2001.4422 371.4394 2001.4449 371.4449 2001.4476 371.4503 2001.4504 371.4557 2001.4531 371.4611 2001.4559 371.4665 2001.4586 371.4719 2001.4613 371.4774 2001.4641 371.4828 2001.4668 371.4882 2001.4695 371.4939 2001.4723 371.4994 2001.475 371.5048 2001.4778 371.5102 2001.4805 371.5157 2001.4832 371.5211 2001.486 371.5265 2001.4887 371.532 2001.4914 371.5374 2001.4942 371.5428 2001.4969 371.5482 2001.4997 371.5536 2001.5024 371.559 2001.5051 371.5644 2001.5079 371.5698 2001.5106 371.5752 2001.5133 371.5806 2001.5161 371.586 2001.5188 371.5914 2001.5216 371.5968 2001.5243 371.6021 2001.527 371.6075 2001.5298 371.6129 2001.5325 371.6182 2001.5352 371.6236 2001.538 371.629 2001.5407 371.6343 2001.5435 371.6396 2001.5462 371.645 2001.5489 371.6503 2001.5517 371.6556 2001.5544 371.661 2001.5572 371.6663 2001.5599 371.6716 2001.5626 371.6769 2001.5654 371.6822 2001.5681 371.6875 2001.5708 371.6928 2001.5736 371.6981 2001.5763 371.7034 2001.5791 371.7086 2001.5818 371.7139 2001.5845 371.7192 2001.5873 371.7244 2001.59 371.7297 2001.5927 371.7377 2001.5955 371.7431 2001.5982 371.7484 2001.601 371.7538 2001.6037 371.7591 2001.6064 371.7644 2001.6092 371.7698 2001.6119 371.7751 2001.6146 371.7804 2001.6174 371.7857 2001.6201 371.791 2001.6229 371.7963 2001.6256 371.8016 2001.6283 371.8069 2001.6311 371.8124 2001.6338 371.8177 2001.6366 371.823 2001.6393 371.8283 2001.642 371.8335 2001.6448 371.8388 2001.6475 371.8441 2001.6502 371.8493 2001.653 371.8546 2001.6557 371.8598 2001.6585 371.8651 2001.6612 371.8703 2001.6639 371.8755 2001.6667 371.8807 2001.6694 371.8861 2001.6721 371.8914 2001.6749 371.8966 2001.6776 371.9018 2001.6804 371.907 2001.6831 371.9122 2001.6858 371.9174 2001.6886 371.9226 2001.6913 371.9277 2001.694 371.9329 2001.6968 371.9381 2001.6995 371.9432 2001.7023 371.9484 2001.705 371.9535 2001.7077 371.9586 2001.7105 371.9637 2001.7132 371.9688 2001.7159 371.9739 2001.7187 371.979 2001.7214 371.9841 2001.7242 371.9892 2001.7269 371.9942 2001.7296 371.9993 2001.7324 372.0044 2001.7351 372.0094 2001.7379 372.0144 2001.7406 372.0195 2001.7433 372.0245 2001.7461 372.0295 2001.7488 372.0345 2001.7515 372.0395 2001.7543 372.0444 2001.757 372.0494 2001.7598 372.0544 2001.7625 372.0593 2001.7652 372.0643 2001.768 372.0692 2001.7707 372.0741 2001.7734 372.079 2001.7762 372.084 2001.7789 372.0889 2001.7817 372.0937 2001.7844 372.0986 2001.7871 372.1035 2001.7899 372.1084 2001.7926 372.1132 2001.7953 372.118 2001.7981 372.1229 2001.8008 372.1277 2001.8036 372.1325 2001.8063 372.1373 2001.809 372.1421 2001.8118 372.1469 2001.8145 372.1517 2001.8172 372.1565 2001.82 372.1612 2001.8227 372.1687 2001.8255 372.1735 2001.8282 372.1783 2001.8309 372.1831 2001.8337 372.1879 2001.8364 372.1927 2001.8392 372.1975 2001.8419 372.2023 2001.8446 372.207 2001.8474 372.2118 2001.8501 372.2165 2001.8528 372.2213 2001.8556 372.226 2001.8583 372.2307 2001.8611 372.2355 2001.8638 372.2402 2001.8665 372.2449 2001.8693 372.2496 2001.872 372.2543 2001.8747 372.259 2001.8775 372.2636 2001.8802 372.2683 2001.883 372.2729 2001.8857 372.2775 2001.8884 372.2821 2001.8912 372.2867 2001.8939 372.2913 2001.8966 372.2959 2001.8994 372.3006 2001.9021 372.3052 2001.9049 372.3098 2001.9076 372.3143 2001.9103 372.3189 2001.9131 372.3234 2001.9158 372.328 2001.9185 372.3325 2001.9213 372.337 2001.924 372.3415 2001.9268 372.346 2001.9295 372.3505 2001.9322 372.3549 2001.935 372.3594 2001.9377 372.3639 2001.9405 372.3684 2001.9432 372.3728 2001.9459 372.3773 2001.9487 372.3817 2001.9514 372.3861 2001.9541 372.3905 2001.9569 372.3949 2001.9596 372.3993 2001.9624 372.4036 2001.9651 372.408 2001.9678 372.4124 2001.9706 372.4167 2001.9733 372.421 2001.976 372.4254 2001.9788 372.4298 2001.9815 372.4341 2001.9843 372.4384 2001.987 372.4427 2001.9897 372.447 2001.9925 372.4512 2001.9952 372.4555 2001.9979 372.4597 2002.0007 372.464 2002.0034 372.4682 2002.0062 372.4724 2002.0089 372.4766 2002.0116 372.4808 2002.0144 372.4851 2002.0171 372.4893 2002.0198 372.4935 2002.0226 372.4977 2002.0253 372.5018 2002.0281 372.506 2002.0308 372.5101 2002.0335 372.5142 2002.0363 372.5184 2002.039 372.5225 2002.0418 372.5266 2002.0445 372.5307 2002.0472 372.5347 2002.05 372.5388 2002.0527 372.5429 2002.0554 372.5469 2002.0582 372.551 2002.0609 372.555 2002.0637 372.559 2002.0664 372.563 2002.0691 372.567 2002.0719 372.571 2002.0746 372.575 2002.0773 372.579 2002.0801 372.5829 2002.0828 372.5869 2002.0856 372.5908 2002.0883 372.5947 2002.091 372.599 2002.0938 372.6029 2002.0965 372.6068 2002.0992 372.6107 2002.102 372.6146 2002.1047 372.6185 2002.1075 372.6224 2002.1102 372.6263 2002.1129 372.6301 2002.1157 372.634 2002.1184 372.6378 2002.1211 372.6416 2002.1239 372.6455 2002.1266 372.6493 2002.1294 372.6531 2002.1321 372.6569 2002.1348 372.6606 2002.1376 372.6644 2002.1403 372.6682 2002.1431 372.6719 2002.1458 372.6757 2002.1485 372.6794 2002.1513 372.6831 2002.154 372.6868 2002.1567 372.6906 2002.1595 372.6942 2002.1622 372.6979 2002.165 372.7016 2002.1677 372.7055 2002.1704 372.7092 2002.1732 372.7128 2002.1759 372.7165 2002.1786 372.7201 2002.1814 372.7238 2002.1841 372.7274 2002.1869 372.731 2002.1896 372.7346 2002.1923 372.7382 2002.1951 372.7418 2002.1978 372.7454 2002.2005 372.749 2002.2033 372.7525 2002.206 372.7561 2002.2088 372.7596 2002.2115 372.7632 2002.2142 372.7667 2002.217 372.7703 2002.2197 372.7738 2002.2225 372.7773 2002.2252 372.7808 2002.2279 372.7843 2002.2307 372.7878 2002.2334 372.7913 2002.2361 372.7947 2002.2389 372.7982 2002.2416 372.8016 2002.2444 372.8051 2002.2471 372.8085 2002.2498 372.8119 2002.2526 372.8153 2002.2553 372.8187 2002.258 372.8221 2002.2608 372.8255 2002.2635 372.8289 2002.2663 372.8323 2002.269 372.8356 2002.2717 372.839 2002.2745 372.8423 2002.2772 372.8457 2002.2799 372.849 2002.2827 372.8525 2002.2854 372.8558 2002.2882 372.8591 2002.2909 372.8624 2002.2936 372.8657 2002.2964 372.869 2002.2991 372.8723 2002.3018 372.8755 2002.3046 372.8788 2002.3073 372.8821 2002.3101 372.8853 2002.3128 372.8886 2002.3155 372.8918 2002.3183 372.8951 2002.321 372.8983 2002.3238 372.9015 2002.3265 372.9047 2002.3292 372.9079 2002.332 372.9111 2002.3347 372.9143 2002.3374 372.9174 2002.3402 372.9206 2002.3429 372.9238 2002.3457 372.9269 2002.3484 372.9301 2002.3511 372.9332 2002.3539 372.9364 2002.3566 372.9395 2002.3593 372.9426 2002.3621 372.9458 2002.3648 372.9489 2002.3676 372.952 2002.3703 372.9551 2002.373 372.9581 2002.3758 372.9612 2002.3785 372.9643 2002.3812 372.9674 2002.384 372.9704 2002.3867 372.9735 2002.3895 372.9765 2002.3922 372.9796 2002.3949 372.9826 2002.3977 372.9856 2002.4004 372.9887 2002.4031 372.9917 2002.4059 372.9947 2002.4086 372.9977 2002.4114 373.0007 2002.4141 373.0037 2002.4168 373.0067 2002.4196 373.0096 2002.4223 373.0126 2002.4251 373.0156 2002.4278 373.0185 2002.4305 373.0215 2002.4333 373.0244 2002.436 373.0274 2002.4387 373.0303 2002.4415 373.0333 2002.4442 373.0362 2002.447 373.0391 2002.4497 373.042 2002.4524 373.0449 2002.4552 373.0478 2002.4579 373.0507 2002.4606 373.0536 2002.4634 373.0565 2002.4661 373.0593 2002.4689 373.0622 2002.4716 373.0651 2002.4743 373.0679 2002.4771 373.0708 2002.4798 373.0736 2002.4825 373.0765 2002.4853 373.0793 2002.488 373.0822 2002.4908 373.085 2002.4935 373.0878 2002.4962 373.0906 2002.499 373.0934 2002.5017 373.0962 2002.5044 373.099 2002.5072 373.1018 2002.5099 373.1046 2002.5127 373.1074 2002.5154 373.1102 2002.5181 373.113 2002.5209 373.1157 2002.5236 373.1185 2002.5264 373.1213 2002.5291 373.124 2002.5318 373.1268 2002.5346 373.1295 2002.5373 373.1323 2002.54 373.135 2002.5428 373.1377 2002.5455 373.1405 2002.5483 373.1432 2002.551 373.1459 2002.5537 373.1486 2002.5565 373.1513 2002.5592 373.154 2002.5619 373.1567 2002.5647 373.1594 2002.5674 373.1621 2002.5702 373.1648 2002.5729 373.1675 2002.5756 373.1702 2002.5784 373.1728 2002.5811 373.1755 2002.5838 373.1782 2002.5866 373.1809 2002.5893 373.1835 2002.5921 373.1862 2002.5948 373.1888 2002.5975 373.1915 2002.6003 373.1941 2002.603 373.1967 2002.6057 373.1994 2002.6085 373.202 2002.6112 373.2046 2002.614 373.2073 2002.6167 373.2099 2002.6194 373.2125 2002.6222 373.2151 2002.6249 373.2177 2002.6277 373.2203 2002.6304 373.2229 2002.6331 373.2255 2002.6359 373.2281 2002.6386 373.2307 2002.6413 373.2333 2002.6441 373.2359 2002.6468 373.2385 2002.6496 373.2411 2002.6523 373.2436 2002.655 373.2462 2002.6578 373.2488 2002.6605 373.2514 2002.6632 373.2539 2002.666 373.2565 2002.6687 373.2591 2002.6715 373.2616 2002.6742 373.2642 2002.6769 373.2668 2002.6797 373.2693 2002.6824 373.2719 2002.6851 373.2744 2002.6879 373.277 2002.6906 373.2795 2002.6934 373.282 2002.6961 373.2846 2002.6988 373.2871 2002.7016 373.2897 2002.7043 373.2921 2002.707 373.2946 2002.7098 373.2972 2002.7125 373.2997 2002.7153 373.3022 2002.718 373.3047 2002.7207 373.3073 2002.7235 373.3098 2002.7262 373.3123 2002.729 373.3148 2002.7317 373.3173 2002.7344 373.3198 2002.7372 373.3223 2002.7399 373.3249 2002.7426 373.3274 2002.7454 373.3299 2002.7481 373.3324 2002.7509 373.3349 2002.7536 373.3374 2002.7563 373.3399 2002.7591 373.3424 2002.7618 373.3449 2002.7645 373.3474 2002.7673 373.3499 2002.77 373.3524 2002.7728 373.3549 2002.7755 373.3573 2002.7782 373.3598 2002.781 373.3623 2002.7837 373.3648 2002.7864 373.3673 2002.7892 373.3698 2002.7919 373.3723 2002.7947 373.3748 2002.7974 373.3773 2002.8001 373.3798 2002.8029 373.3822 2002.8056 373.3847 2002.8084 373.3872 2002.8111 373.3897 2002.8138 373.3922 2002.8166 373.3947 2002.8193 373.3971 2002.822 373.3996 2002.8248 373.4021 ------------------------------ Message: 104 Date: Tue, 23 Nov 2010 11:14:05 +1300 From: David Scott <d.scott at auckland.ac.nz> To: Henri Mone <henriMone at gmail.com> Cc: "r-help at r-project.org" <r-help at r-project.org> Subject: Re: [R] cpgram: access data, confidence bands Message-ID: <4CEAEB2D.3000306 at auckland.ac.nz> Content-Type: text/plain; charset=ISO-8859-1; format=flowed On 22/11/10 22:54, Henri Mone wrote:> Dear R experts, beginners and everyone else, > > I'm calculating "cumulative periodogram" using the command "cpgram" > [1] from the MASS library. Here is a short example with the "lh" > (hormone level) dataset: > > library(MASS) > plot(lh,type="l",ylab="value",xlab="time", main="Hormone Levels (lh)") > spectrum(lh, main="Hormone Levels (lh)") # periodigram > cpgram(lh, main="Hormone Levels (lh)") # cumul. periodigram > > I got following two questions: > > 1. The command "cpgram" plots the cumulative periodogram without any > problem. But I could not figure out any way to access the data of the > plot (save it in a variable). > the following command fails (contains no data): > >myObject<-cpgram(lh, main="Hormone Levels (lh)") > >summary(myObject) > Length Class Mode > 0 NULL NULL > > Is there an easy way to access the data of the cumulative > periodogram, or do I need to rewrite the "cpgram" function? >You need to rewrite cpgram. Have a look at the last line of the function, it is invisible() meaning it doesn't return anything. It is easy to change it, replace the last line by for example return(list(pgram = y, cum = cumsum(y)/sum(y))) or whatever you actually want to return.> 2. The "cpgram" function plots with the default options the 95% > confidence bands in the plot. The confidence band are defined such > that in 95% of the cases the true value will lie inside the bands. For > most cases which I tested the cumulative periodogram is outside the > confidence band. Does "cpgram" plot the confidence band of the the > cumulative periodogram or for the periodogram (I think it is the > cumulative periodigram, is this correct?). How should the confidence > band in "cpgram" be interpreted? Some more description on this would > be great. > >It is the cumulative periodogram (as the name suggests). What did you test? Only a white noise process should stay inside the confidence bands. There is some information about the use of the cumulative periodogram in Venables and Ripley's book for which cpgram was written (but admittedly not a lot). David Scott _________________________________________________________________ David Scott Department of Statistics The University of Auckland, PB 92019 Auckland 1142, NEW ZEALAND Phone: +64 9 923 5055, or +64 9 373 7599 ext 85055 Email: d.scott at auckland.ac.nz, Fax: +64 9 373 7018 Director of Consulting, Department of Statistics ------------------------------ Message: 105 Date: Mon, 22 Nov 2010 17:21:09 -0500 From: David Winsemius <dwinsemius at comcast.net> To: "Kenney, Colleen T CTR USA AMC" <colleen.t.kenney at us.army.mil> Cc: r-help at r-project.org Subject: Re: [R] Probit Analysis: Confidence Interval for the LD50 using Fieller's and Heterogeneity (UNCLASSIFIED) Message-ID: <99B2DFBB-F6B5-4555-BBF7-4D5261612C4E at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Nov 22, 2010, at 2:24 PM, Kenney, Colleen T CTR USA AMC wrote:> Classification: UNCLASSIFIED > Caveats: NONE > > A similar question has been posted in the past but never answered. My > question is this: for probit analysis, how do you program a 95% > confidence interval for the LD50 (or LC50, ec50, etc.), including a > heterogeneity factor as written about in "Probit Analysis" by > Finney(1971)? The heterogeneity factor comes into play through the > chi-squared test for homogeneity and is equal to h=chi^2/(k-2), > where k > is the number of doses and k-2 are the degrees of freedom. > > I have done a lot of research on this and really appreciate any help[[elided Yahoo spam]] The reason it may not have had a rely ... assuming the questions was homomorphic to this one... is that there is no R content in this question. You may want to post on: http://stats.stackexchange.com/ But I suggest you be somewhat more descriptive than just citing a 40 year-old text. -- David Winsemius, MD West Hartford, CT ------------------------------ Message: 106 Date: Mon, 22 Nov 2010 14:28:56 -0800 From: Dennis Murphy <djmuser at gmail.com> To: "Kenney, Colleen T CTR USA AMC" <colleen.t.kenney at us.army.mil> Cc: r-help at r-project.org Subject: Re: [R] Probit Analysis: Confidence Interval for the LD50 using Fieller's and Heterogeneity (UNCLASSIFIED) Message-ID: <AANLkTinRodwVPCrXm8SnfAfcn3QQ6-qJg4GnaNZMG0B3 at mail.gmail.com> Content-Type: text/plain Hi: The MASS package has a function dose.p() to produce a CI for ED50, ED90 or EDp in general (0 < p < 100). It takes a model object (presumably from a suitable logistic regression) as input. You could always take the code already available and adapt it to your situation or you could investigate one or more of the packages devoted to dose-response models. A useful thing to know in R is the sos package, starting with its primary function, findFn(): library(sos) findFn('ED50') scares up the three major packages related to dose-response models with 40 matches to ED50. HTH, Dennis On Mon, Nov 22, 2010 at 11:24 AM, Kenney, Colleen T CTR USA AMC < colleen.t.kenney at us.army.mil> wrote:> Classification: UNCLASSIFIED > Caveats: NONE > > A similar question has been posted in the past but never answered. My > question is this: for probit analysis, how do you program a 95% > confidence interval for the LD50 (or LC50, ec50, etc.), including a > heterogeneity factor as written about in "Probit Analysis" by > Finney(1971)? The heterogeneity factor comes into play through the > chi-squared test for homogeneity and is equal to h=chi^2/(k-2), where k > is the number of doses and k-2 are the degrees of freedom. > > I have done a lot of research on this and really appreciate any help[[elided Yahoo spam]]> > > Classification: UNCLASSIFIED > Caveats: NONE > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. >[[alternative HTML version deleted]] ------------------------------ Message: 107 Date: Mon, 22 Nov 2010 14:36:08 -0800 (PST) To: r-help at r-project.org Subject: [R] empity value in colnames Message-ID: <1290465368687-3054564.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii Hi Guys. I have a matrix which has names in every other column and the other one is empity ("") example X10000 X10001 X10002 [1,] "A" "G" "A" "G" "G" [2,] "G" "G" "A" "G" "A" [3,] "G" "G" "A" "A" "A" [4,] "G" "G" "A" "A" "A" [5,] "A" "G" "A" "A" "A" I am creating another matrix (I called here subset) which cbinds information from another matrix (not shown) and a subset of this example matrix above (let's say column 1 and 2) and save as .prn file subset Sequence family clone female male X10000 V2 X10001 V4 1 40003 400 540003 10005 22055 A G A G 2 40011 400 540011 10005 22055 G G A G 3 40014 400 540014 10005 22055 G G A A 4 40042 400 540042 10005 22055 G G A A 5 40057 400 540057 10005 22055 A G A A Everytime I do it, it creates a column name ("V2" and "V4" in bold) where it should be empty (""). Do you guys have any clue on how to to this? Thanks -- View this message in context: http://r.789695.n4.nabble.com/empity-value-in-colnames-tp3054564p3054564.htm l Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 108 Date: Mon, 22 Nov 2010 19:59:04 -0300 From: Kjetil Halvorsen <kjetilbrinchmannhalvorsen at gmail.com> To: Mike Marchywka <marchywka at hotmail.com> Cc: r-help at r-project.org, ligges at statistik.tu-dortmund.de Subject: Re: [R] ?summaryRprof running at 100% cpu for one hour ... Message-ID: <AANLkTi=nDYVu_cuYuH3_0HOe1T99B_cy3ZA_VrgG_FnE at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 see below. On Mon, Nov 22, 2010 at 12:57 PM, Mike Marchywka <marchywka at hotmail.com> wrote:> > > > > > > > > ---------------------------------------- >> Date: Mon, 22 Nov 2010 12:41:06 -0300 >> Subject: Re: [R] ?summaryRprof running at 100% cpu for one hour ... >> From: kjetilbrinchmannhalvorsen at gmail.com >> To: marchywka at hotmail.com >> CC: ligges at statistik.tu-dortmund.de; r-help at r-project.org >> >> see below. >> >> On Mon, Nov 22, 2010 at 12:13 PM, Mike Marchywka wrote: >> > >> > >> Thanks. Will try. Really, I tried yesterday, to run R under gdb within >> emacs, but it did'nt work out. What I did (in emacs 23) was, typing >> Ctrl-u M-x R >> and then enter the option >> --debugger=gdb >>[[elided Yahoo spam]]>> >> Kjetil > > I rarely use gdb but it did seem to work with R but I executed gdb from > cygwin windoh and IIRC ctrl-C worked fine as it broke into debugger. > I guess you could try that- start gdb and attach or invoke R from gdb. > >OK, thanks. I started R with R --debugger=gdb in a shell, outside emacs. then it works. I did some unsystematic sampling with Ctrl-C. Most of the time it was stuck in memory.c, apparently doing garbage collection. Other files which occured was unique.c, duplicate.c kjetil>------------------------------ Message: 109 Date: Mon, 22 Nov 2010 18:13:17 -0500 From: David Winsemius <dwinsemius at comcast.net> Cc: r-help at r-project.org Subject: Re: [R] empity value in colnames Message-ID: <AB076202-FC9B-40FA-891B-9420CCF73EB8 at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Nov 22, 2010, at 5:36 PM, M.Ribeiro wrote:> > Hi Guys. > I have a matrix which has names in every other column and the other > one is > empity ("") > > example > X10000 X10001 X10002 > [1,] "A" "G" "A" "G" "G" > [2,] "G" "G" "A" "G" "A" > [3,] "G" "G" "A" "A" "A" > [4,] "G" "G" "A" "A" "A" > [5,] "A" "G" "A" "A" "A" > > I am creating another matrixHow?> (I called here subset) which cbinds information > from another matrix (not shown) and a subset of this example matrix > aboveOff hand, I would guess that this "other matrix", of undescribed structure, is really a dataframe and you are actually calling cbind.data.frame> (let's say column 1 and 2) and save as .prn file > > subset > Sequence family clone female male X10000 V2 X10001 V4 > 1 40003 400 540003 10005 22055 A G A G > 2 40011 400 540011 10005 22055 G G A G > 3 40014 400 540014 10005 22055 G G A A > 4 40042 400 540042 10005 22055 G G A A > 5 40057 400 540057 10005 22055 A G A A > > > Everytime I do it, it creates a column name ("V2" and "V4" in bold) > where it > should be empty ("").That's why I asked "How?"> Do you guys have any clue on howExactly. ............. How? You show us your "how" and we will show you ours.> to to this? > > Thanks > >David Winsemius, MD West Hartford, CT ------------------------------ Message: 110 Date: Mon, 22 Nov 2010 18:17:30 -0500 From: Mike Marchywka <marchywka at hotmail.com> To: <kjetilbrinchmannhalvorsen at gmail.com> Cc: r-help at r-project.org, ligges at statistik.tu-dortmund.de Subject: Re: [R] ?summaryRprof running at 100% cpu for one hour ... Message-ID: <BLU113-W16E5F5F3BA4752BD2EDCAEBE3D0 at phx.gbl> Content-Type: text/plain; charset="iso-8859-1" ----------------------------------------> Date: Mon, 22 Nov 2010 19:59:04 -0300 > Subject: Re: [R] ?summaryRprof running at 100% cpu for one hour ... > From: kjetilbrinchmannhalvorsen at gmail.com > To: marchywka at hotmail.com > CC: ligges at statistik.tu-dortmund.de; r-help at r-project.org > > see below. > > On Mon, Nov 22, 2010 at 12:57 PM, Mike Marchywka wrote: > > > > > > > > > > > > > > > > > > ---------------------------------------- > >> Date: Mon, 22 Nov 2010 12:41:06 -0300 > >> Subject: Re: [R] ?summaryRprof running at 100% cpu for one hour ... > >> From: kjetilbrinchmannhalvorsen at gmail.com > >> To: marchywka at hotmail.com > >> CC: ligges at statistik.tu-dortmund.de; r-help at r-project.org > >> > >> see below. > >> > >> On Mon, Nov 22, 2010 at 12:13 PM, Mike Marchywka wrote: > >> > > >> > > >> Thanks. Will try. Really, I tried yesterday, to run R under gdb within > >> emacs, but it did'nt work out. What I did (in emacs 23) was, typing > >> Ctrl-u M-x R > >> and then enter the option > >> --debugger=gdb > >>[[elided Hotmail spam]]> >> > >> Kjetil > > > > I rarely use gdb but it did seem to work with R but I executed gdb from > > cygwin windoh and IIRC ctrl-C worked fine as it broke into debugger. > > I guess you could try that- start gdb and attach or invoke R from gdb. > > > > > > OK, thanks. I started R with > R --debugger=gdb > in a shell, outside emacs. then it works. > > I did some unsystematic sampling with Ctrl-C. Most of the time it wasstuck> in memory.c, apparently doing garbage collection. > Other files which occured was unique.c, duplicate.c >you may want to try the R-develop list for better help now but presumably you can get symobls somewhere and a readable stack trace. I guess floundering with memory management would be consistent with high CPU usage since as far as the OS is concerned the process is runnable. In java you see stuff like this with lots of temp objects being created. I guess if it is gc and you make lots of garbage and then need a big contiguous area could slow things down a lot. Once you are pretty sure you stopped it in a hotspot, you can try stepping in and out of things and see if anything looks odd. I guess one other exploratory thing to try, this may or may not work in R with your problem, is get a snapshot of the memory and then use a utility like "strings" to see if there is any indication of what is going on. If objects are annotated at all something may jump out but hard to know.> kjetil > > > >------------------------------ Message: 111 Date: Mon, 22 Nov 2010 15:18:40 -0800 From: Kendric Wang <kendricw at interchange.ubc.ca> To: r-help at r-project.org Subject: [R] Sporadic errors when training models using CARET Message-ID: <AANLkTimKokkPu6L4r1dmaunSvd3PJ-dAjed_CKZ84RWD at mail.gmail.com> Content-Type: text/plain Hi. I am trying to construct a svmLinear model using the "caret" package (see code below). Using the same data, without changing any setting, sometimes it constructs the model successfully, and sometimes I get an index out of bounds error. Is this unexpected behaviour? I would appreciate any insights this issue. Thanks. ~Kendric> train.y[1] S S S S R R R R R R R R R R R R R R R R R R R R Levels: R S> train.xm1 m2 1 0.1756 0.6502 2 0.1110 -0.2217 3 0.0837 -0.1809 4 -0.3703 -0.2476 5 8.3825 2.8814 6 5.6400 12.9922 7 7.5537 7.4809 8 3.5005 5.7844 9 16.8541 16.6326 10 9.1851 8.7814 11 1.4405 11.0132 12 9.8795 2.6182 13 8.7151 4.5476 14 -0.2092 -0.7601 15 3.6876 2.5772 16 8.3776 5.0882 17 8.6567 7.2640 18 20.9386 20.1107 19 12.2903 4.7864 20 10.5920 7.5204 21 10.2679 9.5493 22 6.2023 11.2333 23 -5.0720 -4.8701 24 6.6417 11.5139> svmLinearGrid <- expand.grid(.C=0.1) > svmLinearFit <- train(train.x, train.y, method="svmLinear",tuneGrid=svmLinearGrid) Fitting: C=0.1 Error in indexes[[j]] : subscript out of bounds> svmLinearFit <- train(train.x, train.y, method="svmLinear",tuneGrid=svmLinearGrid) Fitting: C=0.1 maximum number of iterations reached 0.0005031579 0.0005026807maximum number of iterations reached 0.0002505857 0.0002506714Error in indexes[[j]] : subscript out of bounds> svmLinearFit <- train(train.x, train.y, method="svmLinear",tuneGrid=svmLinearGrid) Fitting: C=0.1 maximum number of iterations reached 0.0003270061 0.0003269764maximum number of iterations reached 7.887867e-05 7.866367e-05maximum number of iterations reached 0.0004087571 0.0004087466Aggregating results Selecting tuning parameters Fitting model on full training set R version 2.11.1 (2010-05-31) x86_64-redhat-linux-gnu locale: [1] LC_CTYPE=en_US.UTF-8 LC_NUMERIC=C [3] LC_TIME=en_US.UTF-8 LC_COLLATE=en_US.UTF-8 [5] LC_MONETARY=C LC_MESSAGES=en_US.UTF-8 [7] LC_PAPER=en_US.UTF-8 LC_NAME=C [9] LC_ADDRESS=C LC_TELEPHONE=C [11] LC_MEASUREMENT=en_US.UTF-8 LC_IDENTIFICATION=C attached base packages: [1] splines stats graphics grDevices utils datasets methods [8] base other attached packages: [1] kernlab_0.9-12 pamr_1.47 survival_2.35-8 cluster_1.12.3 [5] e1071_1.5-24 class_7.3-2 caret_4.70 reshape_0.8.3 [9] plyr_1.2.1 lattice_0.18-8 loaded via a namespace (and not attached): [1] grid_2.11.1 -- MSc. Candidate CIHR/MSFHR Training Program in Bioinformatics University of British Columbia [[alternative HTML version deleted]] ------------------------------ Message: 112 Date: Mon, 22 Nov 2010 15:20:56 -0800 (PST) From: Phil Spector <spector at stat.berkeley.edu> To: madr <madrazel at interia.pl> Cc: r-help at r-project.org Subject: Re: [R] I need a very specific unique like function and I don't know even how to properly call this Message-ID: <alpine.DEB.2.00.1011221517010.14043 at springer.Berkeley.EDU> Content-Type: TEXT/PLAIN; charset=US-ASCII; format=flowed Given a vector, x, we can test if the value above it is equal to itself with abv = c(FALSE,x[-l] == x[-1]) and if the value below is equal to itself with blw = c(x[-l] == x[-1],FALSE) So, for your problem:> abv = c(FALSE,dat[,2][-l] == dat[,2][-1]) > blw = c(dat[,2][-l] == dat[,2][-1],FALSE) > dat[!(abv & blw),][,1] [,2] [1,] 3 7 [2,] 6 5 [3,] 5 5 [4,] 8 4 [5,] 7 4 [6,] 0 6 - Phil Spector Statistical Computing Facility Department of Statistics UC Berkeley spector at stat.berkeley.edu On Mon, 22 Nov 2010, madr wrote:> > consider this matrix: > > [,1] [,2] > [1,] 3 7 > [2,] 6 5 > [3,] 7 5 > [4,] 3 5 > [5,] 7 5 > [6,] 5 5 > [7,] 8 4 > [8,] 2 4 > [9,] 7 4 > [10,] 0 6 > > I need to delete all rows where column 2 above and below has the samevalue,> so the effect would be: > > [,1] [,2] > [1,] 3 7 > [2,] 6 5 > [6,] 5 5 > [7,] 8 4 > [9,] 7 4 > [10,] 0 6 > > is there a built in function for that kind of operation or I must writeone> from scratch ? > Is there a name for that kind of operation ? > -- > View this message in context:http://r.789695.n4.nabble.com/I-need-a-very-specific-unique-like-function-an d-I-don-t-know-even-how-to-properly-call-this-tp3054427p3054427.html> Sent from the R help mailing list archive at Nabble.com. > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >------------------------------ Message: 113 Date: Mon, 22 Nov 2010 18:43:41 -0500 From: David Winsemius <dwinsemius at comcast.net> To: shubha <shuba.pandit at gmail.com> Cc: r-help at r-project.org Subject: Re: [R] how do remove those predictor which have p value greater than 0.05 in GLM? Message-ID: <0911E03A-8277-4731-8AF4-E80777201E28 at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Nov 22, 2010, at 5:10 PM, shubha wrote:> > Thanks for the response, Frank. > I am not saying that I want to delete a variables because of p>0.5.Presumably that was meant to be p > 0.05> But my > concern was: I am using backward stepwise logistic regression, it > keeps the > variables in the final model if the variable significantly > contributing in > the model.Isn't that what backwards selection does?> Otherwise, it should not be in the final model.You're sure? How did you arrive at that conclusion?> Using other software, they give correct results.Correct? Please describe your standards for correctness.> But R, did not. I want > those variables if p<0.05, otherwise exclude from the model.But you said above that was _not_ what you wanted. I'm confused about your posture here.> If you include > that variables, it will affect the Log likelihood ratio and AIC.Yes, perhaps it will, ... so is the standard a p-value or is a penalized penalized estimate? When you toss out a variable, you are deluding yourself to then later ignore that act of deletion when specifying your degrees of freedom for the multiple hypothesis testing effort you have conducting.> I want to > change a P-value criterion <=0.05 in the model. Any suggestions.More reading. Less reliance on canned software.> thanks > > --David Winsemius, MD West Hartford, CT ------------------------------ Message: 114 Date: Mon, 22 Nov 2010 16:22:11 -0800 From: Peter Ehlers <ehlers at ucalgary.ca> To: Marcin Gomulka <mrgomel at gmail.com> Cc: "r-help at r-project.org" <r-help at r-project.org> Subject: Re: [R] plotting a timeline Message-ID: <4CEB0933.30406 at ucalgary.ca> Content-Type: text/plain; charset=ISO-8859-1; format=flowed On 2010-11-20 14:26, Marcin Gomulka wrote:> I was trying to recreate this kind of timeline plot: > http://www.vertex42.com/ExcelArticles/create-a-timeline.html > > As you can see in their excel example, the events are nicely placed out on > both sides of the timeline axis. > > AFAIK there is no function to do this nicely in R-project. Furthermore, > graphics and lattice packages are unable to draw the x-axis in the middleof> the plot. (datapoints cannot be plotted below the axis, as in the Excel > example). > > My question: Is there a function to draw the x-axis inside the plot? (at a > certain vertical position?) Is there a function for the whole timelineplot> that I do not know about? > > I tried to visually replicate the plot using additional elements to create > my own x-axis (code below). I have placed the x-axis ticks on a fixed > y-height (-0.1 and -0.05), but this might look badly with a different > dataset or at other image proportions. I'd rather do this with a dedicated > package function ( like axis() ). >It wouldn't be difficult to write such a function. Here's some code to get you started: with(the_data, { plot(eventtime, impact, type="h", axes=FALSE, ann=FALSE, col="grey", ylim=c(-.7,1.2)) points(eventtime, impact, pch=95, font=5, cex=2, col=4) text(eventtime, impact, label, pos = 1 + 2*(impact > 0)) }) abline(h=0, lwd=2) axis(1, pos=0, lwd=2, lwd.ticks=1) Peter Ehlers> -- > mrgomel > ----------------------- > Below is my example code in R: > > > the_data<- > structure(list(eventtime = c(1914L, 1917L, 1918L, 1939L, 1945L, > 1963L, 1989L, 2001L, 2003L), impact = c(1, -.5, 0.8, 1, 0.8, 0.5, > -.5, 0.5, 1), label = structure(c(8L, 7L, 4L, 9L, 5L, 2L, 3L, > 1L, 6L), .Label = c("9/11", "Cuban crisis", "end of communism", > "end WW1", "end WW2", "Iraq war", "start of communism", "WW1", > "WW2"), class = "factor")), .Names = c("eventtime", "impact", > "label"), class = "data.frame", row.names = c(NA, -9L)) > > > plot(the_data$eventtime, the_data$impact, type="h", frame.plot=FALSE, axes > FALSE, xlab="",ylab="", col="grey") > text(the_data$eventtime,the_data$impact, the_data$label) > #axis(1) > abline(h=0,lwd=2) > text(axTicks(1),-0.1, axTicks(1)) > points(axTicks(1),rep(-0.05,length(axTicks(1))), type="h") > > [[alternative HTML version deleted]] > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 115 Date: Tue, 23 Nov 2010 01:36:47 +0100 From: "Sebastian Rudnick" <rudnick at igb-berlin.de> To: r-help at r-project.org Subject: [R] Gap between graph and axis Message-ID: <20101123002430.M57525 at igb-berlin.de> Content-Type: text/plain; charset=utf-8 Hi everyone! I want to plot some precipitation data via plot(type="h"). Unfortunately there is always a gap between the bars and the x-axis, so that the bars reach in the "negative area" below 0 at the y-axis, which is very misleading. The ylim-parameter is set to 0 and max of precipitation, the min value of precipitation is 0 as well. I tried to fix this via the fig parameter, but I have no idea how to do it at all. I hope anyone can help. Thanks a lot, Sebastian ------------------------------ Message: 116 Date: Tue, 23 Nov 2010 11:52:29 +1100 From: <Bill.Venables at csiro.au> To: <rudnick at igb-berlin.de>, <r-help at r-project.org> Subject: Re: [R] Gap between graph and axis Message-ID: <1BDAE2969943D540934EE8B4EF68F95FB27A44FC0D at EXNSW-MBX03.nexus.csiro. au> Content-Type: text/plain; charset="us-ascii" perhaps you need something like this. par(yaxs = "i") plot(runif(10), type = "h", ylim = c(0, 1.1)) -----Original Message----- From: r-help-bounces at r-project.org [mailto:r-help-bounces at r-project.org] On Behalf Of Sebastian Rudnick Sent: Tuesday, 23 November 2010 10:37 AM To: r-help at r-project.org Subject: [R] Gap between graph and axis Hi everyone! I want to plot some precipitation data via plot(type="h"). Unfortunately there is always a gap between the bars and the x-axis, so that the bars reach in the "negative area" below 0 at the y-axis, which is very misleading. The ylim-parameter is set to 0 and max of precipitation, the min value of precipitation is 0 as well. I tried to fix this via the fig parameter, but I have no idea how to do it at all. I hope anyone can help. Thanks a lot, Sebastian ______________________________________________ R-help at r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. ------------------------------ Message: 117 Date: Mon, 22 Nov 2010 19:01:16 -0600 From: Erin Hodgess <erinm.hodgess at gmail.com> To: R help <r-help at stat.math.ethz.ch> Subject: [R] R on Androids? Message-ID: <AANLkTik5jP2cm5Gy0cDRvbkQURoQVc0vgYmE-hDYWG_B at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Dear R People: Does R run on the Android, yet, please? I'm about 99% sure that it does not, but thought that I would double check. Thanks, Erin -- Erin Hodgess Associate Professor Department of Computer and Mathematical Sciences University of Houston - Downtown mailto: erinm.hodgess at gmail.com ------------------------------ Message: 118 Date: Mon, 22 Nov 2010 20:07:08 -0500 From: watashi at post.com To: yuliya.rmail at gmail.com, r-help at r-project.org Subject: Re: [R] how to loop through variables in R? Message-ID: <8CD58C1E2D65661-1718-A0F at web-mmc-m05.sysops.aol.com> Content-Type: text/plain -----Original Message----- From: Yuliya Matveyeva <yuliya.rmail at gmail.com> To: watashi at post.com Sent: Mon, Nov 22, 2010 4:32 pm Subject: Re: [R] how to loop through variables in R? If you want to have a name-specific loop. Assign names to your variables after inserting them into the data.frame like that: colnames(df) <- c("var1","var23","var456","var44",...) for (nam in colnames(df)) { myfunction(df[[nam]]) } Data.frames support access by names. Unfortunately, in this sense I have to type in 1000 times... isn't there any function that allows retrieving and assigning all columns automatically? read.table in "the R intro manual" just brief says it can read a table but then there's no follow up about how to access the data.... [[alternative HTML version deleted]] ------------------------------ Message: 119 Date: Mon, 22 Nov 2010 17:20:22 -0800 (PST) To: r-help at r-project.org Subject: [R] How to start default browser on R Message-ID: <116352.18459.qm at web113205.mail.gq1.yahoo.com> Content-Type: text/plain; charset=utf-8 Hi folks, Win7 64 bit IE 64 bit How to start IE on R? TIA B.R. Stephen L ------------------------------ Message: 120 Date: Mon, 22 Nov 2010 17:23:29 -0800 (PST) To: R help <r-help at stat.math.ethz.ch> Subject: [R] question on "uniCox" Message-ID: <741278.83327.qm at web30803.mail.mud.yahoo.com> Content-Type: text/plain; charset=utf-8 Hi list, I?m testing out uniCox R package (version 1.0, on R2.12.0, WinXP). When I ran uniCox on my data, there are always some NA?s in the beta matrix, which in turn causes problems in uniCoxCV call. I don?t see anything wrong with the corresponding data (e.g. no NAs) and if I fit a univariate Cox model, the features that give NA beta estimates are actually pretty significant. Could you please let me know what happened and how to avoid this? I?ve attached the outputs of the function calls below. Thank you very much! ...Tao> a <- uniCox(x=t(dat.ave.train.base), y=sampleinfo.ave.train.base$tm2dthr, >status=sampleinfo.ave.train.base$censrdth)lambda value 1 out of 20 lambda value 2 out of 20 lambda value 3 out of 20 lambda value 4 out of 20 lambda value 5 out of 20 lambda value 6 out of 20 lambda value 7 out of 20 lambda value 8 out of 20 lambda value 9 out of 20 lambda value 10 out of 20 lambda value 11 out of 20 lambda value 12 out of 20 lambda value 13 out of 20 lambda value 14 out of 20 lambda value 15 out of 20 lambda value 16 out of 20 lambda value 17 out of 20 lambda value 18 out of 20 lambda value 19 out of 20 lambda value 20 out of 20 5 betas missing> aa <- uniCoxCV(a, x=t(dat.ave.train.base), >y=sampleinfo.ave.train.base$tm2dthr,status=sampleinfo.ave.train.base$censrdth) FOLD= 1 lambda value 1 out of 20 lambda value 2 out of 20 lambda value 3 out of 20 lambda value 4 out of 20 lambda value 5 out of 20 lambda value 6 out of 20 lambda value 7 out of 20 lambda value 8 out of 20 lambda value 9 out of 20 lambda value 10 out of 20 lambda value 11 out of 20 lambda value 12 out of 20 lambda value 13 out of 20 lambda value 14 out of 20 lambda value 15 out of 20 lambda value 16 out of 20 lambda value 17 out of 20 lambda value 18 out of 20 lambda value 19 out of 20 lambda value 20 out of 20 3 betas missing 1 Error in coxph(Surv(y[ii], status[ii]) ~ eta.new) : No (non-missing) observations> a[[2]][(rowSums(is.na(a[[2]])))>0,][,1] [,2] [,3] [,4] [,5] [,6] [,7] [,8] [,9] [,10] [,11] [,12] [,13] [,14] [,15] [,16] [,17] [,18] [,19] [,20] [1,] 92.6641 NaN NaN NaN 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 [2,] NaN 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 [3,] 567.3650 NaN 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ------------------------------ Message: 121 Date: Mon, 22 Nov 2010 17:28:57 -0800 From: Spencer Graves <spencer.graves at structuremonitoring.com> To: Georg Otto <gwo at well.ox.ac.uk> Cc: r-help at stat.math.ethz.ch Subject: Re: [R] Find in R and R books Message-ID: <4CEB18D9.6030305 at structuremonitoring.com> Content-Type: text/plain; charset=ISO-8859-1; format=flowed Other people like R Site Search (http://search.r-project.org/nmz.html), which is available via the standard R function "RSiteSearch". For me, the fastest literature search on virtually anything statistical is the "findFn" function in the "sos" package. (Disclaimer: I'm the lead author of that package, so I may be biased.) "findFn" sorts search results to put the package with the most matches first. The print method opens a table of the results in a web browser with hot links to the individual matches. "sos" comes with a vignette, which includes an example of the "writeFindFn2xls" function. This writes a "findFn" object to an Excel file with two sheets: The second is all the matches found. The first is a summary of the packages found with extra information not available via RSiteSearch. Hope this helps. Spencer On 11/22/2010 3:19 AM, Georg Otto wrote:> > >> Also when I try to search in google using for example the word R insidethe search lemma I get very few results as the R confuses the search engine. When I was looking something in matlab ofcourse it was easier to get results as the search engine performs better.>> What are your tricks when you want to find some function that providessome functionality?> To search R-specific sites the best place to go is this one: > > http://www.rseek.org/ > > Cheers, > > Georg > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >------------------------------ Message: 122 Date: Mon, 22 Nov 2010 17:33:22 -0800 (PST) To: Terry Therneau <therneau at mayo.edu> Cc: r-help at r-project.org, dieter.menne at menne-biomed.de, r_tingley at hotmail.com Subject: Re: [R] calculating martingale residual on new data using "predict.coxph" Message-ID: <616997.31494.qm at web30804.mail.mud.yahoo.com> Content-Type: text/plain; charset=us-ascii Thank you, Terry! ----- Original Message ----> From: Terry Therneau <therneau at mayo.edu>> Cc: r-help at r-project.org; dieter.menne at menne-biomed.de; r_tingley at hotmail.com> Sent: Mon, November 22, 2010 6:11:15 AM > Subject: Re: calculating martingale residual on new data using"predict.coxph"> > This feature has been added in survival 2.36-1, which is now on CRAN. > (2.36-2 should appear in another day or so) > Terry T. > > ---------begin included message -------- > I was trying to use "predict.coxph" to calculate martingale residuals on > a test > data, however, as pointed out before > > http://tolstoy.newcastle.edu.au/R/e4/help/08/06/13508.html > > predict(mycox1, newdata, type="expected") is not implemented yet. > >------------------------------ Message: 123 Date: Mon, 22 Nov 2010 17:36:18 -0800 (PST) To: Frank Harrell <f.harrell at vanderbilt.edu>, r-help at r-project.org Subject: Re: [R] calculating martingale residual on new data using Message-ID: <562346.4353.qm at web30805.mail.mud.yahoo.com> Content-Type: text/plain; charset=us-ascii [[elided Yahoo spam]] ...Tao ----- Original Message ----> From: Frank Harrell <f.harrell at vanderbilt.edu> > To: r-help at r-project.org > Sent: Sun, November 21, 2010 5:49:36 AM > Subject: Re: [R] calculating martingale residual on new data using > > > The tendency is to use residual-like diagnostics on the entire datasetthat> was available for model development. For test data we typically run > predictive accuracy analyses. For example, one of the strongestvalidations> is to show, in a high-resolution calibration plot, that absolutepredictions> (e.g., probability of survival at 2 years) are accurate. > > Frank > > > ----- > Frank Harrell > Department of Biostatistics, Vanderbilt University > -- > View this message in context: >http://r.789695.n4.nabble.com/calculating-martingale-residual-on-new-data-using-predict-coxph-tp3050712p3052377.html> > Sent from the R help mailing list archive at Nabble.com. > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >------------------------------ Message: 124 Date: Mon, 22 Nov 2010 17:36:54 -0800 (PST) To: David Winsemius <dwinsemius at comcast.net> Cc: r-help at r-project.org, dieter.menne at menne-biomed.de, r_tingley at hotmail.com Subject: Re: [R] calculating martingale residual on new data using "predict.coxph" Message-ID: <212275.29986.qm at web30802.mail.mud.yahoo.com> Content-Type: text/plain; charset=us-ascii [[elided Yahoo spam]] ...Tao ----- Original Message ----> From: David Winsemius <dwinsemius at comcast.net>> Cc: r-help at r-project.org; dieter.menne at menne-biomed.de; r_tingley at hotmail.com> Sent: Sun, November 21, 2010 5:50:31 AM > Subject: Re: [R] calculating martingale residual on new data using >"predict.coxph" > > > On Nov 21, 2010, at 3:42 AM, Shi, Tao wrote: > > > Hi David, > > > > Thanks, but I don't quite follow your examples below. > > I wasn't really sure they did anything useful anyway. > > > The residuals you > > calculated are still based on the training data from which your coxmodel>was > > generated. I'm interested in the testing data. > > The survest function in rms and the survfit function in survival will >calculate survival probabilities given a model and newdata, and dependingon>your definition of "residual" you could take the difference between the >calculation and validation data. That must be what happens (at least at agross>level of description) when Harrell runs his validate function on his cphmodels>in the rms/Design package, although I don't know if something that youwould>recognize as a martingale residual is an identifiable intermediate. > > If you are using survfit, it would appear from my reading that youwould>need to set the "individual" parameter to TRUE. I'm assuming you plannedto>calculate these (1- expected) at the event times of the validation cohort >(which it appears the default method fixes via the censor argument)? > > --David > > > > > > > Best, > > > > ...Tao > > > > > > > > > > > > ----- Original Message ---- > >> From: David Winsemius <dwinsemius at comcast.net> > >> To: David Winsemius <dwinsemius at comcast.net>> >> dieter.menne at menne-biomed.de; r_tingley at hotmail.com > >> Sent: Fri, November 19, 2010 10:53:26 AM > >> Subject: Re: [R] calculating martingale residual on new data using > >> "predict.coxph" > >> > >> > >> On Nov 19, 2010, at 12:50 PM, David Winsemius wrote: > >> > >>> > >>> On Nov 19, 2010, at 12:32 PM, Shi, Tao wrote: > >>> > >>>> Hi list, > >>>> > >>>> I was trying to use "predict.coxph" to calculate martingaleresiduals on>a > >> test > >>>> data, however, as pointed out before > >>> > >>> What about resid(fit) ? It's my reading of Therneau & Gramsch [andof> >> help(coxph.object) ] that they consider those martingale residuals. > >> > >> The manner in which I _thought_ this would work was to insert somedummy>cases > >> into the original data and then to get residuals by weighting thecases> >> appropriately. That doesn't seem to be as successful as I imagined: > >> > >>> test1 <- list(time=c(4,3,1,1,2,2,3,3), weights=c(rep(1,7), 0), > >> + status=c(1,1,1,0,1,1,0,1), > >> + x=c(0,2,1,1,1,0,0,1), > >> + sex=c(0,0,0,0,1,1,1,1)) > >>> coxph(Surv(time, status) ~ x , test1, weights=weights)$weights > >> Error in fitter(X, Y, strats, offset, init, control, weights weights, >: > >> Invalid weights, must be >0 > >> # OK then make it a small number > >>> test1 <- list(time=c(4,3,1,1,2,2,3,3), weights=c(rep(1,7), 0.01), > >> + status=c(1,1,1,0,1,1,0,1), > >> + x=c(0,2,1,1,1,0,0,1), > >> + sex=c(0,0,0,0,1,1,1,1)) > >>> print(resid( coxph(Surv(time, status) ~ x , test1,weights=weights))> >> ,digits=3) > >> 1 2 3 4 5 6 7 8 > >> -0.6410 -0.5889 0.8456 -0.1544 0.4862 0.6931 -0.6410 0.0509 > >> Now take out constructed case and weights > >> > >>> test1 <- list(time=c(4,3,1,1,2,2,3), > >> + status=c(1,1,1,0,1,1,0), > >> + x=c(0,2,1,1,1,0,0), > >> + sex=c(0,0,0,0,1,1,1)) > >>> print(resid( coxph(Surv(time, status) ~ x , test1) ) ,digits=3) > >> 1 2 3 4 5 6 7 > >> -0.632 -0.589 0.846 -0.154 0.486 0.676 -0.632 > >> > >> Expecting approximately the same residuals for first 7 cases but not>really > >> getting it. There must be something about weights in coxph that Idon't> >> understand, unless a one-hundreth of a case gets "up indexed" insidethe> >> machinery of coxph? > >> > >> Still think that inserting a single constructed case into a realdataset>of > >> sufficient size ought to be able to yield some sort of estimate, andonly>be a > >> minor perturbation, although I must admit I'm having troublefiguring out>... > >> why are we attempting such a maneuver? The notion of "residuals"around> >> constructed cases makes me statistically suspicious, although Isuppose>that is > >> just some sort of cumulative excess/deficit death fraction. > >> > >>>> http://tolstoy.newcastle.edu.au/R/e4/help/08/06/13508.html > >>>> > >>>> predict(mycox1, newdata, type="expected") is not implemented yet. >Dieter > >>>> suggested to use 'cph' and 'predict.Design', but from my reading sofar,> >> I'm not > >>>> sure they can do that. > >>>> > >>>> Do you other ways to calculate martingale residuals on a new data? > >>>>[[elided Yahoo spam]]> >>>> > >>>> ...Tao > >> > >> --David Winsemius, MD > >> West Hartford, CT > >> > >> > > > > > > > > David Winsemius, MD > West Hartford, CT > >------------------------------ Message: 125 Date: Mon, 22 Nov 2010 20:43:56 -0500 From: Mike Marchywka <marchywka at hotmail.com> To: <spencer.graves at structuremonitoring.com>, <gwo at well.ox.ac.uk> Cc: r-help at stat.math.ethz.ch Subject: Re: [R] Find in R and R books Message-ID: <BLU113-W111F8D13F02BDD7AC3C938BE3E0 at phx.gbl> Content-Type: text/plain; charset="iso-8859-1" ----------------------------------------> Date: Mon, 22 Nov 2010 17:28:57 -0800 > From: spencer.graves at structuremonitoring.com > To: gwo at well.ox.ac.uk > CC: r-help at stat.math.ethz.ch > Subject: Re: [R] Find in R and R books > > Other people like R Site Search > (http://search.r-project.org/nmz.html), which is available via the > standard R function "RSiteSearch". > > > For me, the fastest literature search on virtually anything > statistical is the "findFn" function in the "sos" package. > (Disclaimer: I'm the lead author of that package, so I may be biased.) > "findFn" sorts search results to put the package with the most matches > first. The print method opens a table of the results in a web browserAgain, I have in past taken docs for various things like R, rendered html to text and used things like grep and built my own indicies. However, your facility does seem in that line of thought. Personally I haven't had a problem with google scholar or probably even citeseer would return good hits, R is not a common english word so I think google can make use of it.> with hot links to the individual matches. "sos" comes with a vignette, > which includes an example of the "writeFindFn2xls" function. This > writes a "findFn" object to an Excel file with two sheets: The second > is all the matches found. The first is a summary of the packages found > with extra information not available via RSiteSearch. > > > Hope this helps. > Spencer > > > On 11/22/2010 3:19 AM, Georg Otto wrote: > > Alaios writes: > > > > > >> Also when I try to search in google using for example the word R insidethe search lemma I get very few results as the R confuses the search engine. When I was looking something in matlab ofcourse it was easier to get results as the search engine performs better.> >> What are your tricks when you want to find some function that providessome functionality?> > To search R-specific sites the best place to go is this one: > > > > http://www.rseek.org/ > > > > Cheers, > > > > Georg > > > > ______________________________________________ > > R-help at r-project.org mailing list > > https://stat.ethz.ch/mailman/listinfo/r-help > > PLEASE do read the posting guide http://www.R-project.org/posting-guide.html> > and provide commented, minimal, self-contained, reproducible code. > > > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 126 Date: Tue, 23 Nov 2010 01:50:28 +0000 (UTC) From: Ben Bolker <bbolker at gmail.com> To: r-help at stat.math.ethz.ch Subject: Re: [R] Fast Two-Dimensional Optimization Message-ID: <loom.20101123T024136-231 at post.gmane.org> Content-Type: text/plain; charset=us-ascii Wonsang You <you <at> ifn-magdeburg.de> writes:> I have attempted "optim" function to solve a two-dimensional optimization > problem. It took around 25 second to complete the procedure. > However, I want to reduce the computation time: less than 7 second. Isthere> any optimization function in R which is very rapid?This is not nearly enough information for us to help. The answer depends on the characteristics of your objective function. You may be able to quadruple the speed of your optimization by coding your objective function more efficiently in R or by re-coding it in C or C++. You may be able to choose better starting conditions. You may be able to pick an optimization method that is more suitable for your objective function (see ?optim and the "optimx" package on r-forge). Ben Bolker ------------------------------ Message: 127 Date: Mon, 22 Nov 2010 21:11:14 -0500 From: watashi at post.com To: r-help at r-project.org Subject: Re: [R] how to loop through variables in R? Message-ID: <8CD58CAD70284B2-1718-DC0 at web-mmc-m05.sysops.aol.com> Content-Type: text/plain d<-read.table("D:\\Working\\Statics.txt") df <- cbind("Q1", "Q2", "Q3", "Q4", "Q5", "Q5A", "Q5B", "Q5C", "Q5D", "Q5E", "Q5F", "Q5G", "Q6", "Q6A", "Q6B", "Q6C", "Q6D", "Q6E", "Q6F", "Q7", "Q8", "Q9") #Than you can loop through them simply by doing: result <- numeric(length(df)) for (i in 1:(length(df)-1)) { result <- chisq.test(table(df[[i]], df[[i+1]])) } and then this error comes out: Error: unexpected '}' in "}" and how can I redirect the output of the chi-square test to a file instead of console output? [[alternative HTML version deleted]] ------------------------------ Message: 128 Date: Tue, 23 Nov 2010 15:16:04 +1300 From: David Scott <d.scott at auckland.ac.nz> Cc: "r-help at r-project.org" <r-help at r-project.org> Subject: Re: [R] How to start default browser on R Message-ID: <4CEB23E4.8090209 at auckland.ac.nz> Content-Type: text/plain; charset=ISO-8859-1; format=flowed On 23/11/10 14:20, Stephen Liu wrote:> Hi folks, > > Win7 64 bit > IE 64 bit > > How to start IE on R? TIA > > B.R. > Stephen L > >?browseURL -- _________________________________________________________________ David Scott Department of Statistics The University of Auckland, PB 92019 Auckland 1142, NEW ZEALAND Phone: +64 9 923 5055, or +64 9 373 7599 ext 85055 Email: d.scott at auckland.ac.nz, Fax: +64 9 373 7018 Director of Consulting, Department of Statistics ------------------------------ Message: 129 Date: Mon, 22 Nov 2010 18:20:35 -0800 From: Spencer Graves <spencer.graves at structuremonitoring.com> To: Mike Marchywka <marchywka at hotmail.com> Cc: r-help at stat.math.ethz.ch, gwo at well.ox.ac.uk Subject: Re: [R] Find in R and R books Message-ID: <4CEB24F3.3040108 at structuremonitoring.com> Content-Type: text/plain; charset=ISO-8859-1; format=flowed Hi, Mike, et al.: <in line> On 11/22/2010 5:43 PM, Mike Marchywka wrote:> > > > > > ---------------------------------------- >> Date: Mon, 22 Nov 2010 17:28:57 -0800 >> From: spencer.graves at structuremonitoring.com >> To: gwo at well.ox.ac.uk >> CC: r-help at stat.math.ethz.ch >> Subject: Re: [R] Find in R and R books >> >> Other people like R Site Search >> (http://search.r-project.org/nmz.html), which is available via the >> standard R function "RSiteSearch". >> >> >> For me, the fastest literature search on virtually anything >> statistical is the "findFn" function in the "sos" package. >> (Disclaimer: I'm the lead author of that package, so I may be biased.) >> "findFn" sorts search results to put the package with the most matches >> first. The print method opens a table of the results in a web browser > Again, I have in past taken docs for various things like R, rendered > html to text and used things like grep and built my own indicies. > However, your facility does seem in that line of thought. > > Personally I haven't had a problem with google scholar or > probably even citeseer would return good hits, R is not > a common english word so I think google can make use of it.Thanks for this. For me, anything that mixes math with worked examples is vastly superior to either alone, because I no longer have to puzzle sometimes for hours over a single line or page of mathematics: I can try a variety of examples and walk through the code line by line until I understand. In that way, I find R packages much more intelligible than theoretical treatises. This is especially true when the R package comes with a vignette or companion documentation with script files working the examples (e.g., like the "scripts" subdirectories for "nmle" and "fda"). Spencer>> with hot links to the individual matches. "sos" comes with a vignette, >> which includes an example of the "writeFindFn2xls" function. This >> writes a "findFn" object to an Excel file with two sheets: The second >> is all the matches found. The first is a summary of the packages found >> with extra information not available via RSiteSearch. >> >> >> Hope this helps. >> Spencer >> >> >> On 11/22/2010 3:19 AM, Georg Otto wrote: >>> Alaios writes: >>> >>> >>>> Also when I try to search in google using for example the word R insidethe search lemma I get very few results as the R confuses the search engine. When I was looking something in matlab ofcourse it was easier to get results as the search engine performs better.>>>> What are your tricks when you want to find some function that providessome functionality?>>> To search R-specific sites the best place to go is this one: >>> >>> http://www.rseek.org/ >>> >>> Cheers, >>> >>> Georg >>> >>> ______________________________________________ >>> R-help at r-project.org mailing list >>> https://stat.ethz.ch/mailman/listinfo/r-help >>> PLEASE do read the posting guide http://www.R-project.org/posting-guide.html>>> and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 130 Date: Mon, 22 Nov 2010 17:38:06 -0800 (PST) From: joeponzio <joe.ponzio at gmail.com> To: r-help at r-project.org Subject: [R] using the "apply" method for functions with multiple inputs Message-ID: <1290476286318-3054719.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii hello r users, i'm trying to use the apply method on a function with several inputs, but cannot figure out how to send multiple arguments to the function (not multiple runs of the same function, but one run of the function including two variables - each used within the function). a <- c(1:10,999,999,999) b <- c(11:20,999,999,999) tfun <- function(x,y){ if( (x = 1 & y !=999) || (x > 1 & x < 999 & y == 999) ) x1 <- 1 else x1 <-0 } #this doesn't work - gives an error " 'y' is missing tfilt <- sapply(data.frame(a,b), tfun) thanks, joe -- View this message in context: http://r.789695.n4.nabble.com/using-the-apply-method-for-functions-with-mult iple-inputs-tp3054719p3054719.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 131 Date: Tue, 23 Nov 2010 11:05:23 +1000 From: Kere Klein <k.klein at uq.edu.au> To: "r-help at R-project.org" <r-help at R-project.org> Subject: [R] permalgorithm Message-ID: <DFD404226A5BC247866AEAE43341B65428367A2CCB at UQEXMB01.soe.uq.edu.au> Content-Type: text/plain; charset="us-ascii" To whom may it concern, What is a proper way to analyse a dataset generated from permalgorithm with time varying covariates? It seems like to me that interval censoring would be the choice but Surv suggests that the event would be regarded as occurred at the end of the time interval... Best wishes, Kere Kerenaftali Klein PhD| Biostatistician | Queensland Clinical Trials & Biostatistics Centre The University of Queensland | School of Population Health | Building 33, Level 1| Princess Alexandra Hospital |Ipswich Road | Woolloongabba QLD 4102 | Australia Ph: +61 7 3176 3062| Fax: +61 7 3176 6826 | Email: k.klein at uq.edu.au | Web: http://www.sph.uq.edu.au/qctbc ------------------------------ Message: 132 Date: Mon, 22 Nov 2010 21:43:25 -0500 From: David Winsemius <dwinsemius at comcast.net> To: joeponzio <joe.ponzio at gmail.com> Cc: r-help at r-project.org Subject: Re: [R] using the "apply" method for functions with multiple inputs Message-ID: <DA7675BA-50F5-4CB0-9610-AF60012C4905 at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Nov 22, 2010, at 8:38 PM, joeponzio wrote:> > hello r users, > > i'm trying to use the apply method on a function with several > inputs, but > cannot figure out how to send multiple arguments to the function (not > multiple runs of the same function, but one run of the function > including > two variables - each used within the function). > > a <- c(1:10,999,999,999) > b <- c(11:20,999,999,999) > > tfun <- function(x,y){ > if( (x = 1 & y !=999) || (x > 1 & x < 999 & y == 999) )There is a problem with that first logical test, assigns 1 to x rather than testing.> x1 <- 1 > else > x1 <-0 > } >> mapply("tfun", a, b) [1] 1 1 1 1 1 1 1 1 1 1 0 0 0 (but see above if that were not what you expected.)> #this doesn't work - gives an error " 'y' is missing > tfilt <- sapply(data.frame(a,b), tfun) > > thanks, > joe > > -- > View this message in context:http://r.789695.n4.nabble.com/using-the-apply-method-for-functions-with-mult iple-inputs-tp3054719p3054719.html> Sent from the R help mailing list archive at Nabble.com. > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.David Winsemius, MD West Hartford, CT ------------------------------ Message: 133 Date: Mon, 22 Nov 2010 19:26:14 -0800 (PST) To: David Scott <d.scott at auckland.ac.nz> Cc: "r-help at r-project.org" <r-help at r-project.org> Subject: Re: [R] How to start default browser on R Message-ID: <389534.66648.qm at web113215.mail.gq1.yahoo.com> Content-Type: text/plain; charset=utf-8 Hi David, Thanks for your advice. According to the Example on ?browseURL I tried: 1) browseURL("file:http://www.r-project.org", browser="C:/Program Files/Internet Explorer/iexplore.exe") It starts a small windows asking for permission to accept ActiveX -> OK IE doesn't start 2) browseURL("file:http://d:/R/R-2.5.1/html/index.html", browser="C:/Program Files/Internet Explorer/iexplore.exe") same result as 1) above What I have missed? TIA B.R. Stephen L ----- Original Message ---- From: David Scott <d.scott at auckland.ac.nz> Cc: "r-help at r-project.org" <r-help at r-project.org> Sent: Tue, November 23, 2010 10:16:04 AM Subject: Re: [R] How to start default browser on R On 23/11/10 14:20, Stephen Liu wrote:> Hi folks, > > Win7 64 bit > IE 64 bit > > How to start IE on R? TIA > > B.R. > Stephen L > >?browseURL -- _________________________________________________________________ David Scott Department of Statistics The University of Auckland, PB 92019 Auckland 1142, NEW ZEALAND Phone: +64 9 923 5055, or +64 9 373 7599 ext 85055 Email: d.scott at auckland.ac.nz, Fax: +64 9 373 7018 Director of Consulting, Department of Statistics ------------------------------ Message: 134 Date: Mon, 22 Nov 2010 22:39:15 -0500 From: Ista Zahn <izahn at psych.rochester.edu> Cc: "r-help at r-project.org" <r-help at r-project.org> Subject: Re: [R] How to start default browser on R Message-ID: <AANLkTik4RKiUwPPDj+TYimm3pdrdxCJ2eoSsiR-0zaaA at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Hi Stephen, I'm not sure if this is the problem, but you almost certainly do not want the "file:" part. Try browseURL("http://www.r-project.org") -Ista> Hi David, > > Thanks for your advice. > > According to the Example on ?browseURL I tried: > > 1) > browseURL("file:http://www.r-project.org", browser="C:/ProgramFiles/Internet> Explorer/iexplore.exe") > > It starts a small windows asking for permission to accept ActiveX > -> OK > > IE doesn't start > > 2) > browseURL("file:http://d:/R/R-2.5.1/html/index.html", browser="C:/Program > Files/Internet Explorer/iexplore.exe") > > same result as 1) above > > > What I have missed? ?TIA > > > B.R. > Stephen L > > > > > ----- Original Message ---- > From: David Scott <d.scott at auckland.ac.nz>> Cc: "r-help at r-project.org" <r-help at r-project.org> > Sent: Tue, November 23, 2010 10:16:04 AM > Subject: Re: [R] How to start default browser on R > > ?On 23/11/10 14:20, Stephen Liu wrote: >> Hi folks, >> >> Win7 64 bit >> IE 64 bit >> >> How to start IE on R? ?TIA >> >> B.R. >> Stephen L >> >> > ?browseURL > > -- > _________________________________________________________________ > David Scott ? ?Department of Statistics > ? ? ? ?The University of Auckland, PB 92019 > ? ? ? ?Auckland 1142, ? ?NEW ZEALAND > Phone: +64 9 923 5055, or +64 9 373 7599 ext 85055 > Email: ? ?d.scott at auckland.ac.nz, ?Fax: +64 9 373 7018 > > Director of Consulting, Department of Statistics > > > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >-- Ista Zahn Graduate student University of Rochester Department of Clinical and Social Psychology http://yourpsyche.org ------------------------------ Message: 135 Date: Mon, 22 Nov 2010 22:49:15 -0500 From: Ravi Varadhan <rvaradhan at jhmi.edu> To: Yogesh Tiwari <yogesh.mpi at googlemail.com> Cc: r-help <r-help at stat.math.ethz.ch> Subject: Re: [R] how to calculate derivative Message-ID: <718097d9ea7a.4ceaf36b at johnshopkins.edu> Content-Type: text/plain; CHARSET=US-ASCII Here is a simple approach: data <- read.table("test-data.txt") deriv <- diff(data$V2) / diff(data$V1) times <- (data$V1[-1] + data$V1[-3545])/2 plot(times, deriv, type="l") Another approach is to smooth the original data and then obtain derivatives from the smooth Ravi. ____________________________________________________________________ Ravi Varadhan, Ph.D. Assistant Professor, Division of Geriatric Medicine and Gerontology School of Medicine Johns Hopkins University Ph. (410) 502-2619 email: rvaradhan at jhmi.edu ----- Original Message ----- From: Yogesh Tiwari <yogesh.mpi at googlemail.com> Date: Monday, November 22, 2010 5:14 pm Subject: [R] how to calculate derivative To: r-help <r-help at stat.math.ethz.ch>> Dear R Users, > > I have trend of two time series of CO2 each 10 years of data. One is > varying > weekly and another is bi-weekly. I want to calculate Growth rate ppmv > / year > of these CO2 trends. Therefore I want to calculate time derivative > ppmv / > year. > > How to do it in R? > > Here I attached example data file, I would appreciate if any one > kindly can > help on it. > > Thanks, > > Regards, > Yogesh > ______________________________________________ > R-help at r-project.org mailing list > > PLEASE do read the posting guide > and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 136 Date: Mon, 22 Nov 2010 19:59:02 -0800 (PST) To: Ista Zahn <izahn at psych.rochester.edu> Cc: "r-help at r-project.org" <r-help at r-project.org> Subject: Re: [R] How to start default browser on R Message-ID: <64349.64091.qm at web113204.mail.gq1.yahoo.com> Content-Type: text/plain; charset=utf-8 Hi Ista, I see. Your advice works. Thanks even without: browser="C:/Program Files/Internet Explorer/iexplore.exe") For non default browser it needs; browser="C:/Program Files/Mozilla Firefoe/firefox.exe" What will be "file:" used for? B.R. Stephen L ----- Original Message ---- From: Ista Zahn <izahn at psych.rochester.edu> Cc: David Scott <d.scott at auckland.ac.nz>; "r-help at r-project.org" <r-help at r-project.org> Sent: Tue, November 23, 2010 11:39:15 AM Subject: Re: [R] How to start default browser on R Hi Stephen, I'm not sure if this is the problem, but you almost certainly do not want the "file:" part. Try browseURL("http://www.r-project.org") -Ista> Hi David, > > Thanks for your advice. > > According to the Example on ?browseURL I tried: > > 1) > browseURL("file:http://www.r-project.org", browser="C:/ProgramFiles/Internet> Explorer/iexplore.exe") > > It starts a small windows asking for permission to accept ActiveX > -> OK > > IE doesn't start > > 2) > browseURL("file:http://d:/R/R-2.5.1/html/index.html", browser="C:/Program > Files/Internet Explorer/iexplore.exe") > > same result as 1) above > > > What I have missed? TIA > > > B.R. > Stephen L > > > > > ----- Original Message ---- > From: David Scott <d.scott at auckland.ac.nz>> Cc: "r-help at r-project.org" <r-help at r-project.org> > Sent: Tue, November 23, 2010 10:16:04 AM > Subject: Re: [R] How to start default browser on R > > On 23/11/10 14:20, Stephen Liu wrote: >> Hi folks, >> >> Win7 64 bit >> IE 64 bit >> >> How to start IE on R? TIA >> >> B.R. >> Stephen L >> >> > ?browseURL > > -- > _________________________________________________________________ > David Scott Department of Statistics > The University of Auckland, PB 92019 > Auckland 1142, NEW ZEALAND > Phone: +64 9 923 5055, or +64 9 373 7599 ext 85055 > Email: d.scott at auckland.ac.nz, Fax: +64 9 373 7018 > > Director of Consulting, Department of Statistics > > > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >-- Ista Zahn Graduate student University of Rochester Department of Clinical and Social Psychology http://yourpsyche.org ------------------------------ Message: 137 Date: Tue, 23 Nov 2010 17:00:36 +1300 From: David Scott <d.scott at auckland.ac.nz> Cc: "r-help at r-project.org" <r-help at r-project.org>, Ista Zahn <izahn at psych.rochester.edu> Subject: Re: [R] How to start default browser on R Message-ID: <4CEB3C64.3000106 at auckland.ac.nz> Content-Type: text/plain; charset=UTF-8; format=flowed On 23/11/10 16:59, Stephen Liu wrote:> Hi Ista, > > I see. Your advice works. Thanks > > even without: > browser="C:/Program Files/Internet Explorer/iexplore.exe") > > For non default browser it needs; > browser="C:/Program Files/Mozilla Firefoe/firefox.exe" > > > What will be "file:" used for? > > > B.R. > Stephen L >You can use it to open a local file on your machine as well. I use this all the time with hwriter which writes html reports. David Scott -- _________________________________________________________________ David Scott Department of Statistics The University of Auckland, PB 92019 Auckland 1142, NEW ZEALAND Phone: +64 9 923 5055, or +64 9 373 7599 ext 85055 Email: d.scott at auckland.ac.nz, Fax: +64 9 373 7018 Director of Consulting, Department of Statistics ------------------------------ Message: 138 Date: Mon, 22 Nov 2010 20:23:29 -0800 From: Spencer Graves <spencer.graves at structuremonitoring.com> To: Ravi Varadhan <rvaradhan at jhmi.edu> Cc: Yogesh Tiwari <yogesh.mpi at googlemail.com>, r-help <r-help at stat.math.ethz.ch> Subject: Re: [R] how to calculate derivative Message-ID: <4CEB41C1.8070106 at structuremonitoring.com> Content-Type: text/plain; charset=ISO-8859-1; format=flowed The "fda" package includes various implementations for smoothing data and differentiating the smooth, per Ravi's alternative approach. This is generally preferable to using first differences of raw data, because differencing raw data amplifies noise, while appropriate smooths eliminate much of the noise, leaving you with what you most want. Ramsay and Silverman (2005) Functional Data Analysis, 2nd ed. (Springer) suggest that if you want a second derivative, it is often wise to use quintic splines, because then the second derivative are cubic splines. (The first derivative of a spline of order k is a spline of order k-1.) An example is provided in Figure 1.2 of Ramsay, Hooker and Graves (2009) Functional Data Analysis with R and Matlab (Springer). However, you don't have to get to book to see that example. Just work through the script file "fdarm-ch01.R" in system.file('scripts', package='fda') on any computer with the 'fda' package installed. Hope this helps. Spencer On 11/22/2010 7:49 PM, Ravi Varadhan wrote:> Here is a simple approach: > > data<- read.table("test-data.txt") > > deriv<- diff(data$V2) / diff(data$V1) > > times<- (data$V1[-1] + data$V1[-3545])/2 > > plot(times, deriv, type="l") > > Another approach is to smooth the original data and then obtainderivatives from the smooth> > Ravi. > > ____________________________________________________________________ > > Ravi Varadhan, Ph.D. > Assistant Professor, > Division of Geriatric Medicine and Gerontology > School of Medicine > Johns Hopkins University > > Ph. (410) 502-2619 > email: rvaradhan at jhmi.edu > > > ----- Original Message ----- > From: Yogesh Tiwari<yogesh.mpi at googlemail.com> > Date: Monday, November 22, 2010 5:14 pm > Subject: [R] how to calculate derivative > To: r-help<r-help at stat.math.ethz.ch> > > >> Dear R Users, >> >> I have trend of two time series of CO2 each 10 years of data. One is >> varying >> weekly and another is bi-weekly. I want to calculate Growth rate ppmv >> / year >> of these CO2 trends. Therefore I want to calculate time derivative >> ppmv / >> year. >> >> How to do it in R? >> >> Here I attached example data file, I would appreciate if any one >> kindly can >> help on it. >> >> Thanks, >> >> Regards, >> Yogesh >> ______________________________________________ >> R-help at r-project.org mailing list >> >> PLEASE do read the posting guide >> and provide commented, minimal, self-contained, reproducible code. > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >-- Spencer Graves, PE, PhD President and Chief Operating Officer Structure Inspection and Monitoring, Inc. 751 Emerson Ct. San Jos?, CA 95126 ph: 408-655-4567 ------------------------------ Message: 139 Date: Mon, 22 Nov 2010 20:29:11 -0800 (PST) From: David Stoffer <dsstoffer at gmail.com> To: r-help at r-project.org Subject: Re: [R] Is it possible to make a matrix to start at row 0? Message-ID: <1290486551690-3054829.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii bogdanno-2 wrote:> > I want to make the matrix to be indexed from row (column) 0, not 1 > Can I do that? How? > Thanks > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. > >Try the Oarray package... it did have some problems at one time, but maybe it has been updated since then. -- View this message in context: http://r.789695.n4.nabble.com/Is-it-possible-to-make-a-matrix-to-start-at-ro w-0-tp3054248p3054829.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 140 Date: Mon, 22 Nov 2010 20:41:05 -0800 From: Steve Bellan <sbellan at berkeley.edu> To: r-help at r-project.org Subject: [R] overlay histograms on map at map coordinates Message-ID: <0B1FE6E7-2E3A-4AD0-9199-A1CB88393D10 at berkeley.edu> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes Hi all, I'm trying to visualize animal movement data characteristics spatiotemporally by overlaying many different histograms on a map. I want the histograms to be plotted at coordinates in the map that matches a region they describe. If I was just doing this once, I'd fiddle in Illustrator after R and avoid the headache. But I'll be doing it many many times so it seems worth the elegance & repeatability. For example, say we have the following data: covariate <- rep(letters[1:10], each = 100) shape.pars <- rep(seq(.5,.7, length.out = 10), each = 100) zz <- rgamma(1000, shape = shape.pars, scale = 2) map <- data.frame(factor = unique(covariate), xx = rnorm(10), yy = rnorm(10)) I'd like to be able to have 10 histograms plotted for each level of the covariate at the coordinates specified in the map data.frame. Ideally, I'd like to be able to specify whether they were plotted to the left, right, top, bottom or center of the coordinates similar to in text() and other graphical functions. Looking around the archives, all I've come across are the viewport system in library(grid) which doesn't seem to be able to handle plot() hist() or other such functions. Otherwise, I could create viewports for each histogram inside the map's coordinate system. Any ideas? Thanks! Steve Steve Bellan MPH, Epidemiology PhD Candidate, Environmental Science, Policy & Management Getz Lab University of California, Berkeley ------------------------------ Message: 141 Date: Mon, 22 Nov 2010 21:35:11 -0800 (PST) From: David Stoffer <dsstoffer at gmail.com> To: r-help at r-project.org Subject: Re: [R] Kalman filter Message-ID: <1290490511301-3054858.post at n4.nabble.com> Content-Type: text/plain; charset=UTF-8 It sounds like you've looked at the DLM, DSE, and SSPIR packages. If not, then certainly check them out. Also, we have code for filtering, smoothing and estimation in our text- go to www.stat.pitt.edu/stoffer/tsa3/ and look at the code for chapter 6. There's not a package for the text, but all the code is in a compressed file that you can download. The examples are discussed in detail in the text, but I think looking at the code (and Appendix R on the site) will be sufficient to set up your problem. David Garten Stuhl wrote:> > Hello, > > > > I have completed my kalman filter problem with more details. > > > > The transition- and the measurement equation is given by > > > > x[t]=A[t]*x[t-1]+B[t]*epsilon[t] > > y[t]=C[t]*x[t]+eta[t] > > > > A, y, B and C are Matrices. Y[t] is the data input vector with 800 > elements > (every t has one element) > > > > My Model is described by the following > (discretisation<http://www.dict.cc/englisch-deutsch/discretisation.html>) > stochastic differential equation > > > > Lambda[t]=lambda[t-1]+kappa*lambda[t]*delta_t+epsilon_l > > R[t]=R[t-1]+mu*delta_t+epsilon_r > > epsilon_l=sigma_l*sqroot(delta_t) > > epsilon_r=sigma_r*sqroot(delta_t) > > > > Ln(S[t])=lambda[t]+R[t] > > > > The paramters for estimation are: > > kappa > > mu > > sigma_l > > sigma_r > > > > The state-space-model for this problem is: > > > > x[t]=(lambda[t], R[t])? > > A[t]=(1-kappa+delta_t, 0; 0, 1+mu) > > B[t]=(1,0;0,1) > > epsilon[t]=(epsilon_l, epsilon_r)? > > C[t]=(1,1) > > Eta[t]=0 > > > > I used serveral alternative methods (dlm, kalmanLike, fkf, kfilter) for > parameter estimation but I don?t understand the syntax and the correct > input > for model estimation. > > > > Can anybody help me, which packed is the most best for my problem and how > is > it to control? > > > > Thanks for helping. > > > > Best, > > Thomas > > [[alternative HTML version deleted]] > > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. > >-- View this message in context: http://r.789695.n4.nabble.com/Kalman-filter-tp3049591p3054858.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 142 Date: Mon, 22 Nov 2010 23:36:53 -0600 From: Mari Pesek <marifrances at gmail.com> To: r-help at r-project.org Subject: [R] (no subject) Message-ID: <AANLkTikd0ymrDTzJfjad=2CFJvLZEmZKP4gXDRO_ctGX at mail.gmail.com> Content-Type: text/plain Dear R Help - I am analyzing data from an ecological experiment and am having problems with the ANOVA functions I've tried thus far. The experiment consists of a blocked/split-plot design, with plant biomass as the response. The following is an overview of the treatments applied (nitrogen addition, phosphorus addition, and seeded/not seeded) and at what level (block, main-plot, and sub-plot): - 6 experimental blocks divided into 2 main-plots - each main-plot divided into 8 sub-plots, for a total of 96 sub-plots (6 blocks * 2 main-plots * 8 sub-plots) - 16 experimental treatment conditions in a 4 x 2 x 2 factorial design: - N at 4 levels for each sub-plot - P at 2 levels for each sub-plot - Seed at two levels for each block (one level per main-plot) - response variable = biomass Block Main Plot 0 : No Seed Sub-Plot 0.1 : N0 P0 Sub-Plot 0.2 : N0 P1 Sub-Plot 0.3 : N1 P0 Sub-Plot 0.4 : N1 P1 Sub-Plot 0.5 : N2 P0 Sub-Plot 0.6 : N2 P1 Sub-Plot 0.7 : N3 P0 Sub-Plot 0.8 : N3 P1 Main Plot 1 : Seed Sub-Plot 1.1 : N0 P0 Sub-Plot 1.2 : N0 P1 Sub-Plot 1.3 : N1 P0 Sub-Plot 1.4 : N1 P1 Sub-Plot 1.5 : N2 P0 Sub-Plot 1.6 : N2 P1 Sub-Plot 1.7 : N3 P0 Sub-Plot 1.8 : N3 P1 I've tried several different approaches to run an ANOVA (lmer, aov, lme) on this data, trying to use type III SSs and include a random factor, but am having problems. Any suggestions? [[elided Yahoo spam]] -Mari Pesek -- Dept of Ecology and Evolutionary Biology University of Kansas Haworth Hall, 1200 Sunnyside Ave Lawrence, KS 66045 [[alternative HTML version deleted]] ------------------------------ Message: 143 Date: Mon, 22 Nov 2010 23:39:17 -0600 From: Mari Pesek <marifrances at gmail.com> To: r-help at r-project.org Subject: [R] factorial ANOVA for block/split-plot design Message-ID: <AANLkTimenMxEoq_WGb6hD_Qifp1Y02dN+T3nh55N8LXP at mail.gmail.com> Content-Type: text/plain Dear R Help - I am analyzing data from an ecological experiment and am having problems with the ANOVA functions I've tried thus far. The experiment consists of a blocked/split-plot design, with plant biomass as the response. The following is an overview of the treatments applied (nitrogen addition, phosphorus addition, and seeded/not seeded) and at what level (block, main-plot, and sub-plot): - 6 experimental blocks divided into 2 main-plots - each main-plot divided into 8 sub-plots, for a total of 96 sub-plots (6 blocks * 2 main-plots * 8 sub-plots) - 16 experimental treatment conditions in a 4 x 2 x 2 factorial design: - N at 4 levels for each sub-plot - P at 2 levels for each sub-plot - Seed at two levels for each block (one level per main-plot) - response variable = biomass Block Main Plot 0 : No Seed Sub-Plot 0.1 : N0 P0 Sub-Plot 0.2 : N0 P1 Sub-Plot 0.3 : N1 P0 Sub-Plot 0.4 : N1 P1 Sub-Plot 0.5 : N2 P0 Sub-Plot 0.6 : N2 P1 Sub-Plot 0.7 : N3 P0 Sub-Plot 0.8 : N3 P1 Main Plot 1 : Seed Sub-Plot 1.1 : N0 P0 Sub-Plot 1.2 : N0 P1 Sub-Plot 1.3 : N1 P0 Sub-Plot 1.4 : N1 P1 Sub-Plot 1.5 : N2 P0 Sub-Plot 1.6 : N2 P1 Sub-Plot 1.7 : N3 P0 Sub-Plot 1.8 : N3 P1 I've tried several different approaches to run an ANOVA (lmer, aov, lme) on this data, trying to use type III SSs and include a random factor, but am having problems. Any suggestions? [[elided Yahoo spam]] -Mari Pesek -- Dept of Ecology and Evolutionary Biology University of Kansas Haworth Hall, 1200 Sunnyside Ave Lawrence, KS 66045 [[alternative HTML version deleted]] ------------------------------ Message: 144 Date: Tue, 23 Nov 2010 00:59:09 -0500 From: "RICHARD M. HEIBERGER" <rmh at temple.edu> To: Mari Pesek <marifrances at gmail.com> Cc: r-help at r-project.org Subject: Re: [R] factorial ANOVA for block/split-plot design Message-ID: <AANLkTimAK7kGT3Lgko+seYu9nKbBcSrnJ7oz4CpA71ib at mail.gmail.com> Content-Type: text/plain Please see the maiz example at the end of ?HH::MMC You will need to all the way to the end of the example. If you don't yet have HH, you can get it with install.packages("HH") If you need to write back to the list, please include your attempts. Rich [[alternative HTML version deleted]] ------------------------------ Message: 145 Date: Mon, 22 Nov 2010 21:25:37 -0800 (PST) From: "dhacademic at gmail.com" <dhacademic at gmail.com> To: r-help at r-project.org Subject: Re: [R] question about constraint minimization Message-ID: <AANLkTim2AgXLOWcaJ7AbFpO9PACz6=TdNJh3nfZ3iie7 at mail.gmail.com> Content-Type: text/plain Dear Prof. Ravi Varadhan, Many thanks for the reply. In my case, besides x1=x3, x1=x4 ("x1=x3=x4" was used in last post), another constraint is needed, x2+x3+x4+...+x12=1.5. So there are 9 variables. I have downloaded your code, but I even do not know how to call your code in R program. Actually, I know very little about R. I spent lots of time to read the R help files as well as lots of post on line, and finally prepared the input file that I pasted in my last post. Unfortunately, it does not work well. Can you please help to revise the input file that can work by using the constrOptim function? Or can you plese show me how to call your code in R and send me the input file? [[elided Yahoo spam]] Best, Hao On Mon, Nov 22, 2010 at 2:31 PM, Ravi Varadhan [via R] < ml-node+3054297-1984476990-202837 at n4.nabble.com<ml-node%2B3054297-1984476990 -202837 at n4.nabble.com>> wrote:> I do not understand the constraint x1 = x3 = x4. If this is correct, you > only have 10 unknown parameters. > > If you can correctly formulate your problem, you can have a look at the > packages "alabama" or "BB". The function `auglag' in "alabama" or the > function `spg' in "BB" may be useful. > > Ravi. > > ------------------------------------------------------- > Ravi Varadhan, Ph.D. > Assistant Professor, > Division of Geriatric Medicine and Gerontology School of Medicine Johns > Hopkins University > > Ph. (410) 502-2619 > email: [hiddenemail]<http://user/SendEmail.jtp?type=node&node=3054297&i=0>> > > -----Original Message----- > From: [hiddenemail]<http://user/SendEmail.jtp?type=node&node=3054297&i=1>[mailto:[hidden> email] <http://user/SendEmail.jtp?type=node&node=3054297&i=2>] On > Behalf Of [hiddenemail]<http://user/SendEmail.jtp?type=node&node=3054297&i=3>> Sent: Monday, November 22, 2010 11:10 AM > To: [hidden email] <http://user/SendEmail.jtp?type=node&node=3054297&i=4> > Subject: Re: [R] question about constraint minimization > > > Hi, > > I have struggled on this "bound optimization with equality constraint" by > using optim function for two days, but still fail to prepare a good input. > Can anyone help to prepare the input for my specific case? Many thanks. > > Best, > Hao > > > On Sat, Nov 20, 2010 at 3:17 AM, Hans W Borchers [via R] < > [hidden email]<http://user/SendEmail.jtp?type=node&node=3054297&i=5><ml-node%2B3051338-309 339578-2> > [hidden email] <http://user/SendEmail.jtp?type=node&node=3054297&i=6>> > > wrote: > > > dhacademic <at> gmail.com <dhacademic <at> gmail.com> writes: > > > > > > > > > > > Hi, > > > > > > I am a beginner of R. There is a question about constraint > minimization. > > A > > > function, y=f(x1,x2,x3....x12), needs to be minimized. There are 3 > > > requirements for the minimization: > > > > > > (1) x2+x3+...+x12=1.5 (x1 is excluded); > > > (2) x1=x3=x4; > > > (3) x1, x3 and x5 are in the range of -1~0, respectively. The rest > > variables > > > (x2, x4, x6, x7, ...., x12) are in the range of 0~1, respectively. > > > > > > The "optim" function is used. And part of my input is as follow, where > > > "xx1r" represents the x12: > > > > > > xx1r=1.5-x[2]-x[1]-x[1]-x[3]-x[4]-x[5]-x[6]-x[7]-x[8]-x[9] > > > start=rnorm(9) > > > up=1:9/1:9*1 > > > lo=1:9/1:9*-1 > > > out=optim(start,f,lower=lo,upper=up,method="L-BFGS-B",hessian=TRUE, > > > control=list(trace=6,maxit=1000)) > > > > > > There are two problems in this input. the "up" and "lo" only define a > > range > > > of -1~1 for x1 to x11, which can not meet the requirement (3). In > > addition, > > > there is not any constraint imposed on x12. I have no idea how to > specify > > a > > > matrix that can impose different constraints on individual variablesin> > a > > > > > > function. Any suggestion is highly appreciated. > > > > > > Best, > > > Hao > > > > > > > I don't see any direct need for real 'constraint' optimization here, > > it is a 'bounded' optimization where you are allowed to use > > > > lower <- c(-1,0,-1,0,-1,0,0,0,0,0,0,0) > > upper <- c( 0,1, 0,0, 0,1,1,1,1,1,1,1) > > > > Otherwise, your description is confusing: > > (1) Did you change f to a new function with 9 variables, eliminating > > x3, x4, and x12 ? > > (2) x4 (being equal to x1) has to be in [-1, 0] but also in [0, 1]? > > (3) If you need to restrict x12 to [0, 1] also, you cannot eliminate > it. > > Either keep x12 and use an equality constraint, or use inequality > > constraints on xxlr. > > > > Hans Werner > > > > ______________________________________________ > > [hidden email] > <http://user/SendEmail.jtp?type=node&node=3051338&i=0>mailing list > > https://stat.ethz.ch/mailman/listinfo/r-help > > PLEASE do read the posting guide > >http://www.R-project.org/posting-guide.html<http://www.r-project.org/posting -guide.html>> > and provide commented, minimal, self-contained, reproducible code. > > > > > > ------------------------------ > > View message @ > > > >http://r.789695.n4.nabble.com/question-about-constraint-minimization-tp30508 <http://r.789695.n4.nabble.com/question-about-constraint-minimization-tp3050 8?by-user=t>> 80p3051338.html > > > > To unsubscribe from question about constraint minimization, click > here< >http://r.789695.n4.nabble.com/template/NamlServlet.jtp?macro=unsubscrib<http ://r.789695.n4.nabble.com/template/NamlServlet.jtp?macro=unsubscrib&by-usert>>e_by_code&node=3050880&code=ZGhhY2FkZW1pY0BnbWFpbC5jb218MzA1MDg4MHwtNjM2Nzc0> > NA==>. > > > > > > -- > View this message in context: > >http://r.789695.n4.nabble.com/question-about-constraint-minimization-tp30508 <http://r.789695.n4.nabble.com/question-about-constraint-minimization-tp3050 8?by-user=t>> 80p3053912.html > Sent from the R help mailing list archive at Nabble.com. > > [[alternative HTML version deleted]] > > ______________________________________________ > [hidden email]<http://user/SendEmail.jtp?type=node&node=3054297&i=7>mailing list> https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide >http://www.R-project.org/posting-guide.html<http://www.r-project.org/posting -guide.html>> and provide commented, minimal, self-contained, reproducible code. > > ______________________________________________ > [hidden email]<http://user/SendEmail.jtp?type=node&node=3054297&i=8>mailing list> https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide >http://www.R-project.org/posting-guide.html<http://www.r-project.org/posting -guide.html>> and provide commented, minimal, self-contained, reproducible code. > > > ------------------------------ > View message @ >http://r.789695.n4.nabble.com/question-about-constraint-minimization-tp30508 80p3054297.html> > To unsubscribe from question about constraint minimization, clickhere<http://r.789695.n4.nabble.com/template/NamlServlet.jtp?macro=unsubscrib e_by_code&node=3050880&code=ZGhhY2FkZW1pY0BnbWFpbC5jb218MzA1MDg4MHwtNjM2Nzc0 NA==>.> >-- View this message in context: http://r.789695.n4.nabble.com/question-about-constraint-minimization-tp30508 80p3054850.html Sent from the R help mailing list archive at Nabble.com. [[alternative HTML version deleted]] ------------------------------ Message: 146 Date: Mon, 22 Nov 2010 21:44:16 -0800 (PST) From: Shai <shainova at gmail.com> To: r-help at r-project.org Subject: [R] Explained GLZ model variation Message-ID: <093c0ad4-6426-48be-aaa5-04993054424a at 26g2000yqv.googlegroups.com> Content-Type: text/plain; charset=ISO-8859-1 Hello, I am using the MASS library to create some GLZ models. How can I know how much (%) of the variation is explained by the model, and if it is significant? I did not see this in the anova() or summary() tables... Thanks! Shai ------------------------------ Message: 147 Date: Mon, 22 Nov 2010 23:11:37 -0700 From: Sarah Berry <escalosobre at gmail.com> To: r-help at r-project.org Subject: [R] Lattice and Quartz Message-ID: <B12499D0-CE15-4C74-9705-BD906D10F956 at gmail.com> Content-Type: text/plain I ran this script in a source file on my Mac: library(lattice) year <- 1900:2000 dollars <- (year-1899)^2 plot(year,dollars) quartz("year") histogram(~dollars) The first plot appears in Quartz 2. The second quartz window, named year, opens but the histogram doesn't appear. However, when I copy and paste this script directly into the R console, both quartz windows (Quartz 2 and year) open and both plots appear. As a counter example, the script below run as a source file, works as expected, and I get two plots in two windows: library(lattice) year <- 1900:2000 dollars <- (year-1899)^2 plot(year,dollars) quartz("year") plot(year, dollars) How do I get the lattice package to generate multiple plots in multiple windows from a script run from a source file? Thank you ahead of time, Sarah B. [[alternative HTML version deleted]] ------------------------------ Message: 148 Date: Mon, 22 Nov 2010 23:20:23 -0700 From: Kayce anderson <kaycelu at gmail.com> To: r-help at r-project.org Subject: [R] compare GLM coefficients Message-ID: <AANLkTi==4cJTdVKgnw1K18KUq-VVXEGyjx70WOGJvMzi at mail.gmail.com> Content-Type: text/plain I have a data set of repeated abundance counts over time. I am investigating whether count data reduced to presence-absence (presence) data will reveal similar population trends. I am using a negative binomial distribution for the glm (package MASS) because the count data contains many zeros and extreme values. "count" and "presence" are annual sums for each metric. I have also included sampling effort (visits) as an independent variable because sampling varies between 29-33 visits per year. My models are: glm.nb(count ~ year + visits) and glm.nb(presence ~ year + visits) I would like to test whether the coefficients for "year" are significantly different between models. Please advise me on the best method to make such a comparison. Thank you, Kayce [[alternative HTML version deleted]] ------------------------------ Message: 149 Date: Tue, 23 Nov 2010 17:55:58 +1100 From: Michael Bedward <michael.bedward at gmail.com> To: Kayce anderson <kaycelu at gmail.com> Cc: r-help at r-project.org Subject: Re: [R] compare GLM coefficients Message-ID: <AANLkTinoxyjWnP6WZMqCGc8Yu5aE0+eAvgfG6ZrXUiby at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Hello Kayce, My (very basic) understanding is that you can't directly compare the coefficients across models that have different response variables, nor could you use AIC and similar metrics of model goodness of fit. Instead, I think you have to carefully define what you mean by "reveal similar population trends". If you treat the model with the count response as your reference, and it predicts (for example) population decline of magnitude X over period T, then you can investigate to what extent this same trend is retrieved by the presence response model. But the specifics of the comparison(s) should be closely tied to the population behaviours / syndromes / critical points that you are most interested in. If there are multiple behaviours of interest you want to know to what extent the presence data perform as well as the count data for each of them. That's my general take on the style of the approach. Hopefully others here will have more detailed and knowledgable comments for you. Michael On 23 November 2010 17:20, Kayce anderson <kaycelu at gmail.com> wrote:> I have a data set of repeated abundance counts over time. ?I am > investigating whether count data reduced to presence-absence (presence)data> will reveal similar population trends. ?I am using a negative binomial > distribution for the glm (package MASS) because the count data containsmany> zeros and extreme values. ?"count" and "presence" are annual sums for each > metric. ?I have also included sampling effort (visits) as an independent > variable because sampling varies between 29-33 visits per year. ?My models > are: > > glm.nb(count ~ year + visits) and > glm.nb(presence ~ year + visits) > > I would like to test whether the coefficients for "year" are significantly > different between models. ?Please advise me on the best method to makesuch> a comparison. > > Thank you, > Kayce > > ? ? ? ?[[alternative HTML version deleted]] > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >------------------------------ Message: 150 Date: Tue, 23 Nov 2010 14:57:23 +0800 From: gireesh bogu <girishbogu at gmail.com> To: R-help at r-project.org Subject: [R] Calculating correlation Message-ID: <AANLkTi=_GXE1192XXSwaCg-WcugVDat4Rd1jKu8_GTzk at mail.gmail.com> Content-Type: text/plain Hi guys I have an input file with multiple columns and and rows. Is it possible to calculate correlation of certain value of certain No (For example x of S1 = 112) with all other values (for example start with x 112 corr a 3 of S1 = x-a 0.2 ) INPUT ******* No S1 S2 S3 S4 Sn a 3 4 45 34 23 x 112 0 12 23 0 b 0 1 23 12 1 n 0 1 0 1 1 OUTPUT *********** No S1 S2 S3 S4 Sn x-a 0.2 0.3 ............... x-x 1 1 ................ x-b 0.......................... x-n 0.9 ....................... [[alternative HTML version deleted]] ------------------------------ Message: 151 Date: Tue, 23 Nov 2010 18:33:59 +1100 From: Jim Lemon <jim at bitwrit.com.au> To: romzero <romzero at yahoo.it> Cc: r-help at r-project.org Subject: Re: [R] Some questione about plot Message-ID: <4CEB6E67.6050608 at bitwrit.com.au> Content-Type: text/plain; charset=ISO-8859-1; format=flowed On 11/22/2010 10:14 PM, romzero wrote:> > Q1: How can i draw LSD (Least significant difference) on a plot? > Like this... > http://r.789695.n4.nabble.com/file/n3053430/LSD.jpg > > Q2: How can i draw the axis secondary scale? >Hi romzero, This is somewhat of a guess, but: segments(2.3,1.6,2.3,1.6+<value of LSD>) text(2.2,1.6+<value of LSD>/2,"LSD(P<=0.05)",adj=1) where <value of LSD> is the length of your line. For the "secondary scale", I assume you mean the little ticks on the right side of the plot. Maybe: plotlim<-par("usr") par(xpd=TRUE) segments(plotlim[2],<y values for ticks>, plotlim[2]+diff(plotlim[1:2])/50),<y values for ticks>) where <y values for ticks> are the vertical positions for the ticks. Jim ------------------------------ Message: 152 Date: Mon, 22 Nov 2010 23:31:55 -0800 (PST) To: David Scott <d.scott at auckland.ac.nz> Cc: "r-help at r-project.org" <r-help at r-project.org>, Ista Zahn <izahn at psych.rochester.edu> Subject: Re: [R] How to start default browser on R Message-ID: <163783.15789.qm at web113203.mail.gq1.yahoo.com> Content-Type: text/plain; charset=utf-8 Hi David, I see. File: gives the full path to the .html file created/download. Then the browser will open that file. Thanks. I don't have hwriter package installed. A side question, what will be the corresponding command on R for check whether a package already installed? similar to; On R (Windows) Packages - Load package B.R. Stephen L ----- Original Message ---- From: David Scott <d.scott at auckland.ac.nz> Cc: Ista Zahn <izahn at psych.rochester.edu>; "r-help at r-project.org" <r-help at r-project.org> Sent: Tue, November 23, 2010 12:00:36 PM Subject: Re: [R] How to start default browser on R On 23/11/10 16:59, Stephen Liu wrote:> Hi Ista, > > I see. Your advice works. Thanks > > even without: > browser="C:/Program Files/Internet Explorer/iexplore.exe") > > For non default browser it needs; > browser="C:/Program Files/Mozilla Firefoe/firefox.exe" > > > What will be "file:" used for? > > > B.R. > Stephen L >You can use it to open a local file on your machine as well. I use this all the time with hwriter which writes html reports. David Scott -- _________________________________________________________________ David Scott Department of Statistics The University of Auckland, PB 92019 Auckland 1142, NEW ZEALAND Phone: +64 9 923 5055, or +64 9 373 7599 ext 85055 Email: d.scott at auckland.ac.nz, Fax: +64 9 373 7018 Director of Consulting, Department of Statistics ------------------------------ Message: 153 Date: Tue, 23 Nov 2010 09:35:57 +0200 From: Tal Galili <tal.galili at gmail.com> To: gireesh bogu <girishbogu at gmail.com> Cc: R-help at r-project.org Subject: Re: [R] Calculating correlation Message-ID: <AANLkTi=VQ9N5a96Ytx1QPJJ65JrmeUCtWeDW+6+DJY=r at mail.gmail.com> Content-Type: text/plain Hi there, I'm not sure I understand your question. What are the two vectors you wish to check their correlation? Are they the two rows x and a? Because from your example it seems you are trying to do a correlation between two singular numbers (so probably I didn't get something straight). Tal ----------------Contact Details:------------------------------------------------------- Contact me: Tal.Galili at gmail.com | 972-52-7275845 Read me: www.talgalili.com (Hebrew) | www.biostatistics.co.il (Hebrew) | www.r-statistics.com (English) ---------------------------------------------------------------------------- ------------------ On Tue, Nov 23, 2010 at 8:57 AM, gireesh bogu <girishbogu at gmail.com> wrote:> Hi guys > > I have an input file with multiple columns and and rows. > Is it possible to calculate correlation of certain value of certain No(For> example x of S1 = 112) with all other values (for example start with x 112 > corr a 3 of S1 = x-a 0.2 ) > > INPUT > ******* > > No S1 S2 S3 S4 Sn > a 3 4 45 34 23 > x 112 0 12 23 0 > b 0 1 23 12 1 > n 0 1 0 1 1 > > OUTPUT > *********** > > No S1 S2 S3 S4 Sn > x-a 0.2 0.3 ............... > x-x 1 1 ................ > x-b 0.......................... > x-n 0.9 ....................... > > [[alternative HTML version deleted]] > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. >[[alternative HTML version deleted]] ------------------------------ Message: 154 Date: Tue, 23 Nov 2010 00:05:15 -0800 (PST) To: r-help at r-project.org Subject: [R] About available datasets on PC Message-ID: <835227.40335.qm at web113207.mail.gq1.yahoo.com> Content-Type: text/plain; charset=utf-8 Hi folks, Win7 On running; data(package = .packages(all.available = TRUE)) it displays a list of datasets under:- Data sets in package ?AER?: But I couldn't call/load all of them on the list.> DJFransesError: object 'DJFranses' not found> CreditCardError: object 'CreditCard' not found But I can call/load;> irisWhether the list shows all available datasets on repo, NOT the datasets already download/installed on PC? If YES how to find the running/available datasets on PC? Packages -> Load Datasets the list looks different. Pls advise. TIA B.R. Stephen L ------------------------------ Message: 155 Date: Tue, 23 Nov 2010 00:24:42 -0800 From: Noah Silverman <noah at smartmediacorp.com> To: r-help <r-help at r-project.org> Subject: [R] More detail in chart axis? Message-ID: <4CEB7A4A.8040305 at smartmediacorp.com> Content-Type: text/plain; charset=ISO-8859-1 Hi, I have a series of data (about 80,000 pairs of x,y). Plotting it shows a great chart. However, R has randomly chosen about 6 labels for my x axis. Now, clearly I can't show them all, but would like some control over the granularity of what is displayed. I can't find anything in the documentation about controlling the axis data labels. Is there a way? Alternately, is there a package that would allow me to zoom into an area of the chart so I can see more detail? Thanks, -N ------------------------------ Message: 156 Date: Tue, 23 Nov 2010 09:26:48 +0100 From: derek eder <derek.eder at lungall.gu.se> To: r-help at r-project.org Subject: [R] Error: cannot allocate vector of size x Gb (64-bit ... yet again) Message-ID: <4CEB7AC8.6090304 at lungall.gu.se> Content-Type: text/plain; charset=ISO-8859-1; format=flowed Hello, I am facing the dreaded "Error: cannot allocate vector of size x Gb" and don't understand enough about R (or operating system) memory management to diagnose and solve the problem -- despite studying previous posts and relevant R help -- e.g.: "Error messages beginning cannot allocate vector of size indicate a failure to obtain memory, either because the size exceeded the address-space limit for a process or, more likely, because the system was unable to provide the memory. [...] On all builds of R, the maximum length (number of elements) of a vector is 2^31 - 1 ~ 2*10^9, as lengths are stored as signed integers. In addition, the storage space cannot exceed the address limit." - from Memory-limits {Base} Simple question: Given 64-bit R (AMD64 Linux) with a ulimit of "unlimited", can the size of an R object exceed the amount of availlable RAM memory? Empirically my system with 4Gb RAM and ample Swap, is failing: > x <- integer(10^9) > object.size(x) 4000000040 bytes > gc() used (Mb) gc trigger (Mb) max used (Mb) Ncells 121195 6.5 350000 18.7 350000 18.7 Vcells 500124024 3815.7 606849099 4629.9 550124408 4197.2 > matrix(x, ncol=16) Error: cannot allocate vector of size 3.7 Gb I don't understand how this operation violates the limits detailed in the Memory-limit help (above). Thank you! Derek Eder ---------------------------------------------------------------------------- --------------------- > version _ platform x86_64-pc-linux-gnu arch x86_64 os linux-gnu system x86_64, linux-gnu status major 2 minor 11.1 year 2010 month 05 day 31 svn rev 52157 language R version.string R version 2.11.1 (2010-05-31) derek at papanca:~$ top top - 09:10:18 up 51 min, 4 users, load average: 0.51, 0.51, 0.45 Tasks: 160 total, 2 running, 158 sleeping, 0 stopped, 0 zombie Cpu(s): 0.0%us, 25.0%sy, 0.0%ni, 75.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st Mem: 3796484k total, 3764852k used, 31632k free, 14204k buffers Swap: 2929660k total, 834240k used, 2095420k free, 94800k cached PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 2854 derek 20 0 239m 9260 5448 S 6 0.2 0:05.53 gnome-terminal 1164 root 20 0 218m 31m 10m S 4 0.8 1:29.71 Xorg 3331 derek 20 0 19276 1324 944 R 1 0.0 0:00.6 top ------------------------------ Message: 157 Date: Tue, 23 Nov 2010 00:41:57 -0800 From: Jeff Newmiller <jdnewmil at dcn.davis.ca.us> Cc: r-help at r-project.org Subject: Re: [R] About available datasets on PC Message-ID: <4CEB7E55.1090209 at dcn.davis.ca.us> Content-Type: text/plain; charset=utf-8; format=flowed You need to load the package and then load the data: library(AER) data("DJFranses") then it will be available for you to work with. Stephen Liu wrote:> Hi folks, > > Win7 > > > On running; > data(package = .packages(all.available = TRUE)) > > it displays a list of datasets under:- > Data sets in package ?AER?: > > But I couldn't call/load all of them on the list. > > >> DJFranses >> > Error: object 'DJFranses' not found > > >> CreditCard >> > Error: object 'CreditCard' not found > > > But I can call/load; > >> iris >> > > Whether the list shows all available datasets on repo, NOT the datasetsalready> download/installed on PC? If YES how to find the running/availabledatasets on> PC? > > > Packages -> Load Datasets > the list looks different. > > > Pls advise. TIA > > B.R. > Stephen L > > > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >------------------------------ Message: 158 Date: Tue, 23 Nov 2010 19:52:47 +1100 From: Jim Lemon <jim at bitwrit.com.au> To: Noah Silverman <noah at smartmediacorp.com> Cc: r-help <r-help at r-project.org> Subject: Re: [R] More detail in chart axis? Message-ID: <4CEB80DF.7090206 at bitwrit.com.au> Content-Type: text/plain; charset=ISO-8859-1; format=flowed On 11/23/2010 07:24 PM, Noah Silverman wrote:> Hi, > > I have a series of data (about 80,000 pairs of x,y). > > Plotting it shows a great chart. However, R has randomly chosen about 6 > labels for my x axis. Now, clearly I can't show them all, but would > like some control over the granularity of what is displayed. I can't > find anything in the documentation about controlling the axis data > labels. Is there a way? > > Alternately, is there a package that would allow me to zoom into an area > of the chart so I can see more detail? >Hi Noah, "axis" will let you choose the positions, but may leave some out if it thinks they are too close. "staxlab" (plotrix) allows you to specify positions and labels and can stagger or rotate the labels. zoomInPlot is only one way to show a section of the plot next to the original. zoomplot (TeachingDemos) allows you to zoom on the same plot. There are other functions that offer different zooming methods. Jim ------------------------------ Message: 159 Date: Tue, 23 Nov 2010 09:47:50 +0100 From: Petr PIKAL <petr.pikal at precheza.cz> To: madr <madrazel at interia.pl> Cc: r-help at r-project.org Subject: Re: [R] how to get rid of unused space on all 4 borders in plot() render Message-ID: <OF5F77406D.9CCFFDAC-ONC12577E4.002F5706-C12577E4.00306366 at precheza. cz> Content-Type: text/plain; charset="US-ASCII" Hi r-help-bounces at r-project.org napsal dne 21.11.2010 19:31:33:> > I have looked into par documentation, and only setting for size of theplot> area was pin. But this setting sets the area as inflexible, that is noR intro manual Chapter 12: Graphical procedures page, 71 A typical figure is....>From that you can find thatpin The current plot dimensions, (width,height), in inches. mar A numerical vector of the form c(bottom, left, top, right) which gives the number of lines of margin to be specified on the four sides of the plot. The default is c(5, 4, 4, 2) + 0.1. So David's remark to look at provided documentation seems to be correct. R is not Tetris but comes with quite good documentation and you shall use it. It would help you a lot. par(mar=c(2,1,1,1)+.1) Regards Petr> matter how I make the window small or big it stays the same. Defaultvalue> has advantage that however it uses plot area that is always smaller than > device area still this area is changing with the window and able to be > bigger. > -- > View this message in context:http://r.789695.n4.nabble.com/how-to-get-rid-of-> unused-space-on-all-4-borders-in-plot-render-tp3052527p3052631.html > Sent from the R help mailing list archive at Nabble.com. > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 160 Date: Tue, 23 Nov 2010 09:56:25 +0100 From: Ivan Calandra <ivan.calandra at uni-hamburg.de> To: r-help at r-project.org Subject: Re: [R] how to loop through variables in R? Message-ID: <4CEB81B9.8050901 at uni-hamburg.de> Content-Type: text/plain; charset=ISO-8859-1; format=flowed Hi! You haven't got any answer probably because you didn't provide a reproducible example. You can do it by copy/pasting the output of dput(d) or dput(df). Moreover, before this email, I couldn't really understand what you were trying to do. It's not crystal clear now, but I think I got it. First, when you read your txt, don't you already have the correct data.frame? What is the difference between d and df? It looks like your cbind() step is complicated. You can also index columns by their index numbers. So let's say you want in df the columns 1 to 5 and 6 to 8 from d. You can do it like this: sel <- c(1:5,6:8) ## creates a vector with the columns indexes you want to have in df df <- d[, sel] ## extract these columns from d and assign it into [[elided Yahoo spam]] You can also so it in one step of course: df <- d[, c(1:5,6:8)] Second, in your loop, you overwrite at each iteration the result from the previous one. You could do something like this: result <- numeric(length(df)) ## shouldn't it be length(df)-1? for (i in 1:(length(df)-1)) { result[i] <- chisq.test(table(df[[i]], df[[i+1]])) ## each computation will be stored in a different element of result } Next, chisq.test() returns a list, so it's not really a good idea to store the output in a vector. Take a look at str(chisq.test(table(df[[1]], df[[2]]))) to know which element(s) you want to keep. You would probably want something like this: chisq.test(table(df[[1]], df[[2]]))[1:3] [[elided Yahoo spam]] result <- vector(mode="list", length=length(df)) ## create a list, shouldn't it here also be length(df)-1? names(result) <- paste("chisq_df[[", 1:length(df), "]]_df[[", (1:length(df))+1, "]]", sep="") ## that way, your list is named, which is easier to remember what is ## what if you have lots of columns for (i in 1:(length(df)-1)) { result[[i]] <- chisq.test(table(df[[i]], df[[i+1]]))[1:3] ## each computation will be stored in a different element of the list } Is it what you're looking for? HTH, Ivan Le 11/23/2010 03:11, watashi at post.com a ?crit :> > d<-read.table("D:\\Working\\Statics.txt") > > df<- cbind("Q1", "Q2", "Q3", "Q4", "Q5", "Q5A", "Q5B", "Q5C", "Q5D","Q5E", "Q5F", "Q5G", "Q6", "Q6A", "Q6B", "Q6C", "Q6D", "Q6E", "Q6F", "Q7", "Q8", "Q9")> #Than you can loop through them simply by doing: > result<- numeric(length(df)) > for (i in 1:(length(df)-1)) { > result<- chisq.test(table(df[[i]], df[[i+1]])) > } > > and then this error comes out: > > Error: unexpected '}' in "}" > > > and how can I redirect the output of the chi-square test to a file insteadof console output?> > > [[alternative HTML version deleted]] > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >-- Ivan CALANDRA PhD Student University of Hamburg Biozentrum Grindel und Zoologisches Museum Abt. S?ugetiere Martin-Luther-King-Platz 3 D-20146 Hamburg, GERMANY +49(0)40 42838 6231 ivan.calandra at uni-hamburg.de ********** http://www.for771.uni-bonn.de http://webapp5.rrz.uni-hamburg.de/mammals/eng/1525_8_1.php ------------------------------ Message: 161 Date: Tue, 23 Nov 2010 10:13:18 +0100 From: Petr PIKAL <petr.pikal at precheza.cz> To: "Ni, Melody Zhifang" <z.ni at imperial.ac.uk> Cc: "'r-help at r-project.org'" <r-help at r-project.org> Subject: [R] Odp: save a regression model that can be used later Message-ID: <OF688B8342.0757530D-ONC12577E4.0032343E-C12577E4.0032B81A at precheza. cz> Content-Type: text/plain; charset="US-ASCII" Hi r-help-bounces at r-project.org napsal dne 22.11.2010 16:02:20:> Hi everyone > > I have a question about how to save a regression model in R and how to > retrieve it for making predictions in a new session. > > To be more specific, I fitted a multilevel logistic regression modelusing the> lmer from the "lme4" package. I then successfully make predictionsusing> fitted(mymodel). > > Since data are complex (three levels, nested, numerous categorical and > continuous data describing types of laparoscopic surgery), the computertakes> quite a while to fit the MLM model. I wonder whether it's possible tosave> the fitted model so that I don't have to fit it again for makingpredictions> every time I start a new R session.When you quit R session with option save =Yes you get a file .Rdata. Whenever you start R with this file you get saved environment back, together with your "mymodel". I recommend for each bigger task to create separate directory in which you can save your .Rhelp and .Rdata file without interfering other tasks. Regards Petr> > I searched the mailing-list archive. Suggestions include using save ()to save> the model as "mymodel.rda" and then use load(mymodel.rda) into theworkspace.> I tried without success (in Windows), returning the error message:"Error in> object$fitted : $ operator is invalid for atomic vectors" > > Did I do anything wrong? Any help on this topic is much appreciated1.11 Data permanency and removing objects from R-intro Regards Petr> > BW, Melody > > -- > Dr Melody Ni > Imperial College > Department of Surgery and Cancer > 10th floor, QEQM Building > St. Mary's Hospital > London W2 1NY > Tel/Fax: +44 (0) 20 331 27657/26309 > z.ni at imperial.ac.uk<mailto:z.ni at imperial.ac.uk> > > [[alternative HTML version deleted]] > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code.------------------------------ Message: 162 Date: Tue, 23 Nov 2010 01:14:03 -0800 From: Jeff Newmiller <jdnewmil at dcn.davis.ca.us> To: Dimitri Shvorob <dimitri.shvorob at gmail.com> Cc: r-help at r-project.org Subject: Re: [R] Lost in POSIX Message-ID: <4CEB85DB.8010203 at dcn.davis.ca.us> Content-Type: text/plain; charset=us-ascii; format=flowed Dimitri Shvorob wrote:>> Nor would I call this much of an improvement in clarity... what about >> > "min"? You want to know the minimum? > > LOL. (And apologies for the insensitivity). Thank you for help, Jeff. This > works, but I am still curious to see a solution based on "trunc", ifanyone> can find it. >You mean like trunc(df$t,units="mins") ? See ?trunc.POSIXt for hints on arguments to "units" parameter... ------------------------------ Message: 163 Date: Tue, 23 Nov 2010 01:21:29 -0800 (PST) To: Jeff Newmiller <jdnewmil at dcn.davis.ca.us> Cc: r-help at r-project.org Subject: Re: [R] About available datasets on PC Message-ID: <976004.93861.qm at web113205.mail.gq1.yahoo.com> Content-Type: text/plain; charset=utf-8 Hi Jeff, Tks for your advice. I got it B.R. satimis ----- Original Message ---- From: Jeff Newmiller <jdnewmil at dcn.davis.ca.us> Cc: r-help at r-project.org Sent: Tue, November 23, 2010 4:41:57 PM Subject: Re: [R] About available datasets on PC You need to load the package and then load the data: library(AER) data("DJFranses") then it will be available for you to work with. Stephen Liu wrote:> Hi folks, > > Win7 > > > On running; > data(package = .packages(all.available = TRUE)) > > it displays a list of datasets under:- > Data sets in package ?AER?: > > But I couldn't call/load all of them on the list. > > >> DJFranses >> > Error: object 'DJFranses' not found > > >> CreditCard >> > Error: object 'CreditCard' not found > > > But I can call/load; > >> iris >> > > Whether the list shows all available datasets on repo, NOT the datasetsalready> > download/installed on PC? If YES how to find the running/availabledatasets on> > PC? > > > Packages -> Load Datasets > the list looks different. > > > Pls advise. TIA > > B.R. > Stephen L > > > > ______________________________________________ > R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guidehttp://www.R-project.org/posting-guide.html> and provide commented, minimal, self-contained, reproducible code. >------------------------------ Message: 164 Date: Tue, 23 Nov 2010 12:16:36 +0200 From: Tal Galili <tal.galili at gmail.com> To: gireesh bogu <girishbogu at gmail.com> Cc: R-help at r-project.org Subject: Re: [R] Calculating correlation Message-ID: <AANLkTi=0fzzzrZ5kCf6thcVyDSzm2+2c=qEGr_pW1rVZ at mail.gmail.com> Content-Type: text/plain Hi Gireesh, I seem to be missing something. Could you give me the formula you think should be used to calculate this? I don't get how you wish to get a "correlation" for only one pair of numbers. (or maybe I didn't understood what you explained - please try a simpler example) ----------------Contact Details:------------------------------------------------------- Contact me: Tal.Galili at gmail.com | 972-52-7275845 Read me: www.talgalili.com (Hebrew) | www.biostatistics.co.il (Hebrew) | www.r-statistics.com (English) ---------------------------------------------------------------------------- ------------------ On Tue, Nov 23, 2010 at 10:03 AM, gireesh bogu <girishbogu at gmail.com> wrote:> in one output > > x - S1 i.e. 112 with all combinations ex: > > > 112 vs 3 [ x-S1 vs a -S1] > > 112 vs 4 [ x-S1 vs a- S2 ] > > 112 vs 45 > > 112 vs 34 > > 112 vs 23 > > 112 vs 112 > > . > > .. > > 112 vs 1 > > > in second output > > x - S2 i.e. 0 with all . ex: > > 0 vs 3 > > 0 vs 4 > > > > > > 0 vs 1 > > > in next output > > x-S3 i.e. 12 with all and so on > > > Probably in the given input I should get 5 outputs because of 5 samples > S1,2,3,4,n. > > If they are more like 69 or some thing then I should get 69 outputs or > everything in one output if possible. > > > Please let me know if it is still confusing. > > > thanx > > On Tue, Nov 23, 2010 at 3:35 PM, Tal Galili <tal.galili at gmail.com> wrote: > >> Hi there, >> I'm not sure I understand your question. >> >> What are the two vectors you wish to check their correlation? >> Are they the two rows x and a? >> Because from your example it seems you are trying to do a correlation >> between two singular numbers (so probably I didn't get somethingstraight).>> >> Tal >> >> >> >> ----------------Contact >> Details:------------------------------------------------------- >> Contact me: Tal.Galili at gmail.com | 972-52-7275845 >> Read me: www.talgalili.com (Hebrew) | www.biostatistics.co.il (Hebrew) | >> www.r-statistics.com (English) >> >>---------------------------------------------------------------------------- ------------------>> >> >> >> >> On Tue, Nov 23, 2010 at 8:57 AM, gireesh bogu<girishbogu at gmail.com>wrote:>> >>> Hi guys >>> >>> I have an input file with multiple columns and and rows. >>> Is it possible to calculate correlation of certain value of certain No >>> (For >>> example x of S1 = 112) with all other values (for example start with x >>> 112 >>> corr a 3 of S1 = x-a 0.2 ) >>> >>> INPUT >>> ******* >>> >>> No S1 S2 S3 S4 Sn >>> a 3 4 45 34 23 >>> x 112 0 12 23 0 >>> b 0 1 23 12 1 >>> n 0 1 0 1 1 >>> >>> OUTPUT >>> *********** >>> >>> No S1 S2 S3 S4 Sn >>> x-a 0.2 0.3 ............... >>> x-x 1 1 ................ >>> x-b 0.......................... >>> x-n 0.9 ....................... >>> >>> [[alternative HTML version deleted]] >>> >>> ______________________________________________ >>> R-help at r-project.org mailing list >>> https://stat.ethz.ch/mailman/listinfo/r-help >>> PLEASE do read the posting guide >>> http://www.R-project.org/posting-guide.html >>> and provide commented, minimal, self-contained, reproducible code. >>> >> >> > > > -- > Regards > > Gireesh > #HP#8375-9256 >[[alternative HTML version deleted]] ------------------------------ _______________________________________________ R-help at r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. End of R-help Digest, Vol 93, Issue 23