나여나
2010-Jul-26 02:35 UTC
[R] After writing data in MMF using SEXP structure, can i reference in R?
Hi all, After writing data in MMF(Memory Map File) using SEXP structure, can i reference in R? If input data is larger than 2GB, Can i reference MMF Data in R? my work environment : R version : 2.11.1 OS : WinXP Pro sp3 Thanks and best regards. Park, Young-Ju from Korea. ---------[ ???????? ???????? ???????? ]---------- ???????? : R-help Digest, Vol 89, Issue 25 ???????? : 2010???? 7???? 25???? ????????????, 19???? 00???? 07???? +0900 ???????????? :r-help-request at r-project.org ???????????? :r-help at r-project.org Send R-help mailing list submissions to [1]r-help at r-project.org To subscribe or unsubscribe via the World Wide Web, visit https://stat.ethz.ch/mailman/listinfo/r-help or, via email, send a message with subject or body 'help' to [2]r-help-request at r-project.org You can reach the person managing the list at [3]r-help-owner at r-project.org When replying, please edit your Subject line so it is more specific than "Re: Contents of R-help digest..." Today's Topics: 1. Re: how to calculate the product of every two elements in two vectors (Dennis Murphy) 2. Re: Question regarding panel data diagnostic (Setlhare Lekgatlhamang) 3. Re: Question regarding panel data diagnostic (Setlhare Lekgatlhamang) 4. Re: Question regarding panel data diagnostic (Setlhare Lekgatlhamang) 5. Re: Trouble retrieving the second largest value from each row of a data.frame (David Winsemius) 6. local polynomial with differnt kernal functions (assaedi76 assaedi76) 7. Weights in mixed models (David R.) 8. Re: Odp: Help me with prediction in linear model (Research student) 9. Re: union data in column (Hadley Wickham) 10. Re: UseR! 2010 - my impressions (Frank E Harrell Jr) 11. Re: , Updating Table (Charles C. Berry) 12. Re: , Updating Table (Duncan Murdoch) 13. Re: glm - prediction of a factor with several levels (Ben Bolker) 14. Doubt about a population competition function (Bruno Bastos Gon?alves) 15. Doubt about a population competition function (Gmail) 16. Book on R's Programming Language (Matt Stati) 17. Re: how to calculate the product of every two elements in two vectors (Henrique Dallazuanna) 18. Re: Book on R's Programming Language (Joshua Wiley) 19. Re: how to calculate the product of every two elements in two vectors (Gabor Grothendieck) 20. Re: Book on R's Programming Language (Joseph Magagnoli) 21. Re: Constrain density to 0 at 0? (Greg Snow) 22. matest function for multiple factors (shabnam k) 23. Re: How to deal with more than 6GB dataset using R? (Greg Snow) 24. Using R to fill ETM+ data gaps? (Abdi, Abdulhakim) 25. How to generate a sequence of dates without hardcoding the year (Felipe Carrillo) 26. Re: How to generate a sequence of dates without hardcoding the year (Henrique Dallazuanna) 27. Re: Trouble retrieving the second largest value from each row of a data.frame ([4]mpward at illinois.edu) 28. Re: (no subject) (Paul Smith) 29. Re: How to generate a sequence of dates without hardcoding the year (jim holtman) 30. Re: glm - prediction of a factor with several levels (zachmohr) 31. Re: Trouble retrieving the second largest value from each row of a data.frame (David Winsemius) 32. Re: Trouble retrieving the second largest value from each row of a data.frame (Joshua Wiley) 33. Re: Trouble retrieving the second largest value from each row of a data.frame (David Winsemius) 34. Re: Trouble retrieving the second largest value from each row of a data.frame (David Winsemius) 35. c-statiscs 95% CI for cox regression model (paaventhan jeyaganth) 36. Re: UseR! 2010 - my impressions (Dirk Eddelbuettel) 37. Re: c-statiscs 95% CI for cox regression model (Frank E Harrell Jr) 38. Equivalent to go-to statement (Michael Haenlein) 39. Outlier Problem in Survreg Function (Vipul Agarwal) 40. Re: Equivalent to go-to statement (Gabor Grothendieck) ---------------------------------------------------------------------- Message: 1 Date: Sat, 24 Jul 2010 03:02:47 -0700 From: Dennis Murphy <[5]djmuser at gmail.com> To: aegea <[6]gcheer3 at gmail.com> Cc: [7]r-help at r-project.org Subject: Re: [R] how to calculate the product of every two elements in two vectors Message-ID: <AANLkTinUJjoCiG47ptPq5Eo_fXYZXuQbUzB=+[8]KLCO6QX at mail.gmail.com> Content-Type: text/plain as.vector(t(outer(A, B))) [1] 9 10 11 12 18 20 22 24 27 30 33 36 HTH, Dennis On Fri, Jul 23, 2010 at 8:11 AM, aegea <[9]gcheer3 at gmail.com> wrote: > > Thanks in advance! > > A=c(1, 2, 3) > B=c (9, 10, 11, 12) > > I want to get C=c(1*9, 1*10, 1*11, 1*12, ....., 3*9, 3*10, 3*11, 3*12)? > C is still a vector with 12 elements > Is there a way to do that? > -- > View this message in context: > [10]http://r.789695.n4.nabble.com/how-to-calculate-the-product-of-every-tw o-elements-in-two-vectors-tp2300299p2300299.html > Sent from the R help mailing list archive at Nabble.com. > > ______________________________________________ > [11]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > [12]http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. > [[alternative HTML version deleted]] ------------------------------ Message: 2 Date: Sat, 24 Jul 2010 12:54:11 +0200 From: "Setlhare Lekgatlhamang" <[13]SetlhareL at bob.bw> To: "amatoallah ouchen" <[14]at.ouchen at gmail.com>, <[15]r-help at r-project.org> Subject: Re: [R] Question regarding panel data diagnostic Message-ID: <[16]25D1D72D6E19D144AB813C9C582E16CF03F7EA27 at BOB-EXCHANGE.bob.bw> Content-Type: text/plain; charset="iso-8859-1" My thought is this: It depends on what you have in the panel. Are your data cross-section data observed over ten years for, say, 3 countries (or regions within the same country)? If so, yes you can perform integration properties (what people usually call unit root test) and then test for cointegration. But if the data are quarterly or monthly, these techniques are not relevant. Hope this helps. Lexi -----Original Message----- From: [17]r-help-bounces at r-project.org [mailto:[18]r-help-bounces at r-project.org] On Behalf Of amatoallah ouchen Sent: Friday, July 23, 2010 12:18 AM To: [19]r-help at r-project.org Subject: [R] Question regarding panel data diagnostic Good day R-listers, I'm currently working on a panel data analysis (N=17, T=5), in order to check for the spurious regression problem, i have to test for stationarity but i've read somewhere that i needn't to test for it as my T<10 , what do you think? if yes is there any other test i have to perform in such case (a kind of cointegration test for small T?) Any hint would be highly appreciated. Ama. * * For searches and help try: * [20]http://www.stata.com/help.cgi?search * [21]http://www.stata.com/support/statalist/faq * [22]http://www.ats.ucla.edu/stat/stata/ ______________________________________________ [23]R-help at r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide [24]http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. DISCLAIMER:\ Sample Disclaimer added in a VBScript.\ ...{{dropped:3}} ------------------------------ Message: 3 Date: Sat, 24 Jul 2010 13:00:05 +0200 From: "Setlhare Lekgatlhamang" <[25]SetlhareL at bob.bw> To: "Setlhare Lekgatlhamang" <[26]SetlhareL at bob.bw>, "amatoallah ouchen" <[27]at.ouchen at gmail.com>, <[28]r-help at r-project.org> Subject: Re: [R] Question regarding panel data diagnostic Message-ID: <[29]25D1D72D6E19D144AB813C9C582E16CF03F7EA28 at BOB-EXCHANGE.bob.bw> Content-Type: text/plain; charset="us-ascii" Let me correct an omission in my response below. The last sentence should read "But if the data are 10 quarterly or monthly values, these techniques are not relevant". Cheers Lexi -----Original Message----- From: [30]r-help-bounces at r-project.org [mailto:[31]r-help-bounces at r-project.org] On Behalf Of Setlhare Lekgatlhamang Sent: Saturday, July 24, 2010 12:54 PM To: amatoallah ouchen; [32]r-help at r-project.org Subject: Re: [R] Question regarding panel data diagnostic My thought is this: It depends on what you have in the panel. Are your data cross-section data observed over ten years for, say, 3 countries (or regions within the same country)? If so, yes you can perform integration properties (what people usually call unit root test) and then test for cointegration. But if the data are quarterly or monthly, these techniques are not relevant. Hope this helps. Lexi -----Original Message----- From: [33]r-help-bounces at r-project.org [mailto:[34]r-help-bounces at r-project.org] On Behalf Of amatoallah ouchen Sent: Friday, July 23, 2010 12:18 AM To: [35]r-help at r-project.org Subject: [R] Question regarding panel data diagnostic Good day R-listers, I'm currently working on a panel data analysis (N=17, T=5), in order to check for the spurious regression problem, i have to test for stationarity but i've read somewhere that i needn't to test for it as my T<10 , what do you think? if yes is there any other test i have to perform in such case (a kind of cointegration test for small T?) Any hint would be highly appreciated. Ama. * * For searches and help try: * [36]http://www.stata.com/help.cgi?search * [37]http://www.stata.com/support/statalist/faq * [38]http://www.ats.ucla.edu/stat/stata/ ______________________________________________ [39]R-help at r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide [40]http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. DISCLAIMER:\ Sample Disclaimer added in a VBScript.\ ......{{dropped:14}} ------------------------------ Message: 4 Date: Sat, 24 Jul 2010 13:01:24 +0200 From: "Setlhare Lekgatlhamang" <[41]SetlhareL at bob.bw> To: "amatoallah ouchen" <[42]at.ouchen at gmail.com>, <[43]r-help at r-project.org> Subject: Re: [R] Question regarding panel data diagnostic Message-ID: <[44]25D1D72D6E19D144AB813C9C582E16CF03F7EA29 at BOB-EXCHANGE.bob.bw> Content-Type: text/plain; charset="us-ascii" Let me correct an omission in my response below. The last sentence should read "But if the data are 10 quarterly or monthly values, these techniques are not relevant". Cheers Lexi -----Original Message----- From: [45]r-help-bounces at r-project.org [mailto:[46]r-help-bounces at r-project.org] On Behalf Of Setlhare Lekgatlhamang Sent: Saturday, July 24, 2010 12:54 PM To: amatoallah ouchen; [47]r-help at r-project.org Subject: Re: [R] Question regarding panel data diagnostic My thought is this: It depends on what you have in the panel. Are your data cross-section data observed over ten years for, say, 3 countries (or regions within the same country)? If so, yes you can perform integration properties (what people usually call unit root test) and then test for cointegration. But if the data are quarterly or monthly, these techniques are not relevant. Hope this helps. Lexi -----Original Message----- From: [48]r-help-bounces at r-project.org [mailto:[49]r-help-bounces at r-project.org] On Behalf Of amatoallah ouchen Sent: Friday, July 23, 2010 12:18 AM To: [50]r-help at r-project.org Subject: [R] Question regarding panel data diagnostic Good day R-listers, I'm currently working on a panel data analysis (N=17, T=5), in order to check for the spurious regression problem, i have to test for stationarity but i've read somewhere that i needn't to test for it as my T<10 , what do you think? if yes is there any other test i have to perform in such case (a kind of cointegration test for small T?) Any hint would be highly appreciated. Ama. * * For searches and help try: * [51]http://www.stata.com/help.cgi?search * [52]http://www.stata.com/support/statalist/faq * [53]http://www.ats.ucla.edu/stat/stata/ ______________________________________________ [54]R-help at r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide [55]http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. DISCLAIMER:\ Sample Disclaimer added in a VBScript.\ ......{{dropped:14}} ------------------------------ Message: 5 Date: Sat, 24 Jul 2010 08:40:05 -0400 From: David Winsemius <[56]dwinsemius at comcast.net> To: <[57]mpward at illinois.edu> Cc: [58]r-help at r-project.org Subject: Re: [R] Trouble retrieving the second largest value from each row of a data.frame Message-ID: <[59]D09340C5-3B64-47FA-A168-8EA347F79747 at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Jul 23, 2010, at 9:20 PM, <[60]mpward at illinois.edu> wrote: > I have a data frame with a couple million lines and want to retrieve > the largest and second largest values in each row, along with the > label of the column these values are in. For example > > row 1 > strongest=-11072 > secondstrongest=-11707 > strongestantenna=value120 > secondstrongantenna=value60 > > Below is the code I am using and a truncated data.frame. Retrieving > the largest value was easy, but I have been getting errors every way > I have tried to retrieve the second largest value. I have not even > tried to retrieve the labels for the value yet. > > Any help would be appreciated > Mike Using Holtman's extract of your data with x as the name and the order function to generate an index to names of the dataframe: > t(apply(x, 1, sort, decreasing=TRUE)[1:3, ]) [, 1] [, 2] [, 3] 1 -11072 -11707 -12471 2 -11176 -11799 -12838 3 -11113 -11778 -12439 4 -11071 -11561 -11653 5 -11067 -11638 -12834 6 -11068 -11698 -12430 7 -11092 -11607 -11709 8 -11061 -11426 -11665 9 -11137 -11736 -12570 10 -11146 -11779 -12537 Putting it all together: matrix( paste( t(apply(x, 1, sort, decreasing=TRUE)[1:3, ]), names(x)[ t(apply(x, 1, order, decreasing=TRUE) [1:3, ]) ]), ncol=3) [, 1] [, 2] [, 3] [1, ] "-11072 value120" "-11707 value60" "-12471 value180" [2, ] "-11176 value120" "-11799 value180" "-12838 value0" [3, ] "-11113 value120" "-11778 value60" "-12439 value180" [4, ] "-11071 value120" "-11561 value240" "-11653 value60" [5, ] "-11067 value120" "-11638 value180" "-12834 value0" [6, ] "-11068 value0" "-11698 value60" "-12430 value120" [7, ] "-11092 value120" "-11607 value240" "-11709 value180" [8, ] "-11061 value120" "-11426 value240" "-11665 value60" [9, ] "-11137 value120" "-11736 value60" "-12570 value180" [10, ] "-11146 value300" "-11779 value0" "-12537 value180" -- David. > > >> data<-data.frame(value0, value60, value120, value180, value240, value300) >> data > value0 value60 value120 value180 value240 value300 > 1 -13007 -11707 -11072 -12471 -12838 -13357 > 2 -12838 -13210 -11176 -11799 -13210 -13845 > 3 -12880 -11778 -11113 -12439 -13089 -13880 > 4 -12805 -11653 -11071 -12385 -11561 -13317 > 5 -12834 -13527 -11067 -11638 -13527 -13873 > 6 -11068 -11698 -12430 -12430 -12430 -12814 > 7 -12807 -14068 -11092 -11709 -11607 -13025 > 8 -12770 -11665 -11061 -12373 -11426 -12805 > 9 -12988 -11736 -11137 -12570 -13467 -13739 > 10 -11779 -12873 -12973 -12537 -12973 -11146 >> #largest value in the row >> strongest<-apply(data, 1, max) >> >> >> #second largest value in the row >> n<-function(data)(1/(min(1/(data[1, ]-max(data[1, ]))))+ >> (max(data[1, ]))) >> secondstrongest<-apply(data, 1, n) > Error in data[1, ] : incorrect number of dimensions >> > > ______________________________________________ > [61]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide [62]http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. ------------------------------ Message: 6 Date: Sat, 24 Jul 2010 04:25:10 -0700 (PDT) From: assaedi76 assaedi76 <[63]assaedi76 at yahoo.com> To: [64]r-help at r-project.org Subject: [R] local polynomial with differnt kernal functions Message-ID: <[65]853644.1608.qm at web45210.mail.sp1.yahoo.com> Content-Type: text/plain Hi, R users ???br> I need to use the function (locpoly) to fit a local poynomial regression model, The defult for kernal function is " normal" , but???I need to use different kernal functions such as :Uniform, Triangular, Epanechnikov, ...... Could someone help me define these functions to fit local polynomial regression model?. Email:[66]assaedi76 at yahoo.com ???br> ???br> Thanks alot [[alternative HTML version deleted]] ------------------------------ Message: 7 Date: Sat, 24 Jul 2010 04:14:44 -0700 (PDT) From: "David R." <[67]drbn at yahoo.com> To: [68]r-help at r-project.org Subject: [R] Weights in mixed models Message-ID: <[69]217686.73973.qm at web113215.mail.gq1.yahoo.com> Content-Type: text/plain; charset=iso-8859-1 Hello everyone, I wonder if sample size?can be used as weight? in a weighted mixed model. Or should I use just the inverse of the variance? For example, the class'lmer' in the 'lme4' package?have an option 'weights' (as in the class 'lm' of 'stats'). In the help of lme4 there is an example using 'herd size' as weight in a mixed model. ? ?So, if I have a measurement data (eg height) of 10 groups (sample size ranging from?30 to 3000 individuals?for each group) can?I??use this number (N, sample size) in the 'weights' option? Is this wrong? ? Finally, what to do if the results (coefficients) of weighing by 'inverse of the variance' or by 'sample size' are very different, even opposite? Thank you very much in advance David ------------------------------ Message: 8 Date: Sat, 24 Jul 2010 02:48:11 -0700 (PDT) From: Research student <[70]vijayamahantesh_s at dell.com> To: [71]r-help at r-project.org Subject: Re: [R] Odp: Help me with prediction in linear model Message-ID: <[72]1279964891930-2300991.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii Thanks Murphy and pikal, I need another help, for fitting first fourier transformation , i used following thing .Please advise on this beer_monthl has 400+ records EXample: > head(beer_monthly) beer 1 93.2 2 96.0 3 95.2 4 77.1 5 70.9 6 64.8 time<-seq(1956, 1995.2, length=length(beer_monthly)) sin.t<-sin(2*pi*time) cos.t<-cos(2*pi*time) beer_fit_fourier=lm(beer_monthly[, 1]~poly(time, 2)+sin.t+cos.t) #this is not working beer_fit_fourier=lm(beer_monthly[, 1]~time+time2+sin.t+cos.t) #it is working #prediction is not workinng tpred_four <- data.frame(time = seq(1995, 1998, length = 20)) predict(beer_fit_fourier, newdata = tpred_four) Is there any way to fit first fourier frequency , Please assist. Thanks in advance -- View this message in context: [73]http://r.789695.n4.nabble.com/Help-me-with-prediction-in-linear-model- tp2297313p2300991.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 9 Date: Sat, 24 Jul 2010 07:53:23 -0500 From: Hadley Wickham <[74]hadley at rice.edu> To: Jeff Newmiller <[75]jdnewmil at dcn.davis.ca.us> Cc: [76]r-help at r-project.org, Fahim Md <[77]fahim.md at gmail.com> Subject: Re: [R] union data in column Message-ID: <AANLkTi=eA4KHr2q7fija+qGbTYnHJPLrLLHgw25+Ki=[78]z at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 On Sat, Jul 24, 2010 at 2:23 AM, Jeff Newmiller <[79]jdnewmil at dcn.davis.ca.us> wrote: > Fahim Md wrote: >> >> Is there any function/way to merge/unite the following data >> >> ?GENEID ? ? ?col1 ? ? ? ? ?col2 ? ? ? ? ? ? col3 ? ? ? ? ? ? ? ?col4 >> ?G234064 ? ? ? ? 1 ? ? ? ? ? ? 0 ? ? ? ? ? ? ? ? ?0 ? ? ? ? ? ? ? ? ? 0 >> ?G234064 ? ? ? ? 1 ? ? ? ? ? ? 0 ? ? ? ? ? ? ? ? ?0 ? ? ? ? ? ? ? ? ? 0 >> ?G234064 ? ? ? ? 1 ? ? ? ? ? ? 0 ? ? ? ? ? ? ? ? ?0 ? ? ? ? ? ? ? ? ? 0 >> ?G234064 ? ? ? ? 0 ? ? ? ? ? ? 1 ? ? ? ? ? ? ? ? ?0 ? ? ? ? ? ? ? ? ? 0 >> ?G234065 ? ? ? ? 0 ? ? ? ? ? ? 1 ? ? ? ? ? ? ? ? ?0 ? ? ? ? ? ? ? ? ? 0 >> ?G234065 ? ? ? ? 0 ? ? ? ? ? ? 1 ? ? ? ? ? ? ? ? ?0 ? ? ? ? ? ? ? ? ? 0 >> ?G234065 ? ? ? ? 0 ? ? ? ? ? ? 1 ? ? ? ? ? ? ? ? ?0 ? ? ? ? ? ? ? ? ? 0 >> ?G234065 ? ? ? ? 0 ? ? ? ? ? ? 0 ? ? ? ? ? ? ? ? ?1 ? ? ? ? ? ? ? ? ? 0 >> ?G234065 ? ? ? ? 0 ? ? ? ? ? ? 0 ? ? ? ? ? ? ? ? ?1 ? ? ? ? ? ? ? ? ? 0 >> ?G234065 ? ? ? ? 0 ? ? ? ? ? ? 0 ? ? ? ? ? ? ? ? ?0 ? ? ? ? ? ? ? ? ? 1 >> >> >> into >> GENEID ? ? ?col1 ? ? ? ? ?col2 ? ? ? ? ? ? col3 ? ? ? ? ? ? ? ?col4 >> ?G234064 ? ? ? ? 1 ? ? ? ? ? ? 1 ? ? ? ? ? ? ? ? ?0 ? ? ? ? ? ? ? ? ? 0 >> // 1 appears in col1 and col2 above, rest are zero >> ?G234065 ? ? ? ? 0 ? ? ? ? ? ? 1 ? ? ? ? ? ? ? ? ?1 ? ? ? ? ? ? ? ? ? 1 >> // 1 appears in col2 , 3 and 4 above. >> >> >> Thank > > Warning on terminology: there is a "merge" function in R that lines up rows > from different tables to make a new set of longer rows (more columns). The > usual term for combining column values from multiple rows is "aggregation". > > In addition to the example offered by Jim Holtzman, here are some other > options in no particular order: > > x <- read.table(textConnection(" GENEID col1 col2 col3 col4 > G234064 1 0 0 0 > G234064 1 0 0 0 > G234064 1 0 0 0 > G234064 0 1 0 0 > G234065 0 1 0 0 > G234065 0 1 0 0 > G234065 0 1 0 0 > G234065 0 0 1 0 > G234065 0 0 1 0 > G234065 0 0 0 1 > "), header=TRUE, as.is=TRUE, row.names=NULL) > closeAllConnections() > > # syntactic repackaging of Jim's basic approach > library(plyr) > ddply( x, .(GENEID), function(df) > {with(as.integer(c(col1=any(col1), col2=any(col2), col3=any(col3), col4=any(col4))))} > ) You can do this a little more succinctly with colwise: any_1 <- function(x) as.integer(any(x)) ddply(x, "GENEID", numcolwise(any_1)) Hadley -- Assistant Professor / Dobelman Family Junior Chair Department of Statistics / Rice University [80]http://had.co.nz/ ------------------------------ Message: 10 Date: Sat, 24 Jul 2010 08:55:01 -0500 From: Frank E Harrell Jr <[81]f.harrell at Vanderbilt.Edu> To: Ravi Varadhan <[82]rvaradhan at jhmi.edu> Cc: "[83]r-help at r-project.org" <[84]r-help at r-project.org> Subject: Re: [R] UseR! 2010 - my impressions Message-ID: <[85]4C4AF0B5.6070300 at vanderbilt.edu> Content-Type: text/plain; charset="ISO-8859-1"; format=flowed On 07/23/2010 06:50 PM, Ravi Varadhan wrote: > Dear UseRs!, > > Everything about UseR! 2010 was terrific! I really mean "everything" - the tutorials, invited talks, kaleidoscope sessions, focus sessions, breakfast, snacks, lunch, conference dinner, shuttle services, and the participants. The organization was fabulous. NIST were gracious hosts, and provided top notch facilities. The rousing speech by Antonio Possolo, who is the chief of Statistical Engineering Division at NIST, set the tempo for the entire conference. Excellent invited lectures by Luke Tierney, Frank Harrell, Mark Handcock, Diethelm Wurtz, Uwe Ligges, and Fritz Leisch. All the sessions that I attended had many interesting ideas and useful contributions. During the whole time that I was there, I could not help but get the feeling that I am a part of something great. > > Before I end, let me add a few words about a special person. This conference would not have been as great as it was without the tireless efforts of Kate Mullen. The great thing about Kate is that she did so much without ever hogging the limelight. Thank you, Kate and thank you NIST! > > I cannot wait for UseR!2011! > > Best, > Ravi. > > ____________________________________________________________________ > > Ravi Varadhan, Ph.D. > Assistant Professor, > Division of Geriatric Medicine and Gerontology > School of Medicine > Johns Hopkins University > > Ph. (410) 502-2619 > email: [86]rvaradhan at jhmi.edu I want to echo what Ravi said. The talks were terrific (thanks to the program committee and the speakers) and Kate Mullen and her team did an extraordinary job in putting the conference together and running it. I am proud to have been a part of it. Thank you all! Frank -- Frank E Harrell Jr Professor and Chairman School of Medicine Department of Biostatistics Vanderbilt University ------------------------------ Message: 11 Date: Sat, 24 Jul 2010 08:25:59 -0700 From: "Charles C. Berry" <[87]cberry at tajo.ucsd.edu> To: Marcus Liu <[88]marcusliu667 at yahoo.com> Cc: [89]r-help at r-project.org Subject: Re: [R] , Updating Table Message-ID: <[90]Pine.LNX.4.64.1007240817250.21422 at tajo.ucsd.edu> Content-Type: text/plain; charset="x-unknown"; Format="flowed" On Fri, 23 Jul 2010, Marcus Liu wrote: > Hi everyone, > > Is there any command for updating table withing a loop?? "Loops? We don't need no stinking loops!" (From 'The Good, the Bad, and the Rgly') tab <- table(data.raw, findInterval(seq(along=data.raw), ind+1 ) ) tab %*% upper.tri(tab, diag=T) or tab2 <- tapply( factor(data.raw), findInterval(seq(along=data.raw), ind+1 ), table) Reduce( "+", tab2, accum=TRUE ) HTH, Chuck p.s. See the posting guide re including a reproducible example with requests like yours. > For instance, at i, I have a table as ZZ = table(data.raw[1:ind[i]]) > where "ind" = c(10, 20, 30, ...).?Then , ZZ will be as follow > > "A" "B" "C" > ?3??? 10?? 2 > > At (i + 1), ZZ = table(data.raw[(ind[i]+1):ind[i+1]]) > > "A" "B" "D" > ?4 ?? 7??? 8 > > Is there any command that can update the table ZZ for each time so that in the above example, ZZ will be > > "A" "B" "C" "D" > ?7??? 17?? 2??? 8 > > Thanks. > > liu > > > > > [[alternative HTML version deleted]] > > Charles C. Berry (858) 534-2098 Dept of Family/Preventive Medicine E mailto:[91]cberry at tajo.ucsd.edu UC San Diego [92]http://famprevmed.ucsd.edu/faculty/cberry/ La Jolla, San Diego 92093-0901 ------------------------------ Message: 12 Date: Sat, 24 Jul 2010 11:51:05 -0400 From: Duncan Murdoch <[93]murdoch.duncan at gmail.com> To: "Charles C. Berry" <[94]cberry at tajo.ucsd.edu> Cc: [95]r-help at r-project.org Subject: Re: [R] , Updating Table Message-ID: <[96]4C4B0BE9.7050409 at gmail.com> Content-Type: text/plain; charset=UTF-8; format=flowed On 24/07/2010 11:25 AM, Charles C. Berry wrote: > On Fri, 23 Jul 2010, Marcus Liu wrote: > >> Hi everyone, >> >> Is there any command for updating table withing a loop?? > > "Loops? We don't need no stinking loops!" > (From 'The Good, the Bad, and the Rgly') Actually, that quote comes from the TreasR of the SieRa MadRe. Duncan Murdoch > tab <- table(data.raw, findInterval(seq(along=data.raw), ind+1 ) ) > tab %*% upper.tri(tab, diag=T) > > or > > tab2 <- tapply( factor(data.raw), findInterval(seq(along=data.raw), ind+1 ), table) > Reduce( "+", tab2, accum=TRUE ) > > HTH, > > Chuck > > p.s. See the posting guide re including a reproducible example with > requests like yours. > >> For instance, at i, I have a table as ZZ = table(data.raw[1:ind[i]]) >> where "ind" = c(10, 20, 30, ...).?Then , ZZ will be as follow >> >> "A" "B" "C" >> ?3??? 10?? 2 >> >> At (i + 1), ZZ = table(data.raw[(ind[i]+1):ind[i+1]]) >> >> "A" "B" "D" >> ?4 ?? 7??? 8 >> >> Is there any command that can update the table ZZ for each time so that in the above example, ZZ will be >> >> "A" "B" "C" "D" >> ?7??? 17?? 2??? 8 >> >> Thanks. >> >> liu >> >> >> >> >> [[alternative HTML version deleted]] >> >> > > Charles C. Berry (858) 534-2098 > Dept of Family/Preventive Medicine > E mailto:[97]cberry at tajo.ucsd.edu UC San Diego > [98]http://famprevmed.ucsd.edu/faculty/cberry/ La Jolla, San Diego 92093-0901 > > > > ------------------------------------------------------------------------ > > ______________________________________________ > [99]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide [100]http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. ------------------------------ Message: 13 Date: Sat, 24 Jul 2010 15:52:52 +0000 (UTC) From: Ben Bolker <[101]bbolker at gmail.com> To: [102]r-help at stat.math.ethz.ch Subject: Re: [R] glm - prediction of a factor with several levels Message-ID: <[103]loom.20100724T175114-259 at post.gmane.org> Content-Type: text/plain; charset=us-ascii blackscorpio <olivier.collignon <at> live.fr> writes: > I'm currently attempting to predict the occurence of an event (factor) > having more than 2 levels with several continuous predictors. The model > being ordinal, I was waiting the glm function to return several intercepts, > which is not the case when looking to my results (I only have one > intercept). I finally managed to perform an ordinal polytomous logisitc > regression with the polr function, which gives several intercepts. > But does anyone know what was the model performed with glm and why only one > intercept was given ? It's not sufficiently clear (to me at least) what you're trying to do. Please provide a minimal reproducible example ... As far as I know, polr is the right way to do ordinal regression; it's not clear how you were trying to use glm to do it. Ben Bolker ------------------------------ Message: 14 Date: Sat, 24 Jul 2010 12:13:30 -0300 From: Bruno Bastos Gon?alves <[104]brubruzao at hotmail.com> To: <[105]r-help at r-project.org> Subject: [R] Doubt about a population competition function Message-ID: <[106]SNT111-DS23835E9F6F65737C7B0BEDBBA40 at phx.gbl> Content-Type: text/plain Hi, I'm doing a function that describe two populations in competition. that's the function that i wrote: exclusao<-function(n10, n20, k1, k2, alfa, beta, t){ n1<-k1-(alfa*n20) n2<-k2-(beta*n10) if(t==0){plot(t, n10, type='b', xlim=range(c(1:t), c (1:t)), ylim=range(n10, n20), xlab='tempo', ylab='tamanho populacional') points(t, n20, type='b', col="red") points(t, n10, type="b", col="black") legend("topleft", c("Pop1", "Pop2"), cex=0.8, col=c ("black", "red"), pch=21:21, lty=1:1); } if(t>0){ for (i in 1:t){ n1[i==1]<-n1 n2[i==1]<-n2 n1[i+1]<-k1-alfa*n2[i] n2[i+1]<-k2-beta*n1[i] if(n1[i]==0){n1[i:t]==0} if(n2[i]==0){n2[i:t]==0} } plot(c(1:t), n1[1:t], type='b', xlim=range(c(1:t), c (1:t)), ylim=range(n1[1:t], n2[1:t]), xlab='tempo', ylab='tamanho populacional') points(c(1:t), n2[1:t], type='b', col="red") legend("topleft", c("Pop1", "Pop2"), cex=0.8, col=c ("black", "red"), pch=21:21, lty=1:1); }} Where n10: size population in time 0, n20: size population in time 0, k1: carrying capacity of the population 1, k2: carrying capacity of the population 2, alfa: competition coefficient of population 2 in population 1, beta: competition coefficient of population 1 in population 2, t: time. and when some population becomes 0 (ZERO), i want that population still 0 (ZERO) until the end of "t". i have tried to put " if(n1[i]==0){n1[i:t]==0} if(n2[i]==0){n2[i:t]==0}" after "n2[i+1]<-k2-beta*n1[i]" in the for function, but nothing happens. What may i do ? Thanks Bruno [[alternative HTML version deleted]] ------------------------------ Message: 15 Date: Sat, 24 Jul 2010 12:15:24 -0300 From: "Gmail" <[107]goncalves.b.b at gmail.com> To: "[108]r-help at r-project.org"@stat.math.ethz.ch Subject: [R] Doubt about a population competition function Message-ID: <822A5DC79A42471AB493F6C4A0817C8E at NotebookBruno> Content-Type: text/plain Hi, I'm doing a function that describe two populations in competition. that's the function that i wrote: exclusao<-function(n10, n20, k1, k2, alfa, beta, t){ n1<-k1-(alfa*n20) n2<-k2-(beta*n10) if(t==0){plot(t, n10, type='b', xlim=range(c(1:t), c (1:t)), ylim=range(n10, n20), xlab='tempo', ylab='tamanho populacional') points(t, n20, type='b', col="red") points(t, n10, type="b", col="black") legend("topleft", c("Pop1", "Pop2"), cex=0.8, col=c ("black", "red"), pch=21:21, lty=1:1); } if(t>0){ for (i in 1:t){ n1[i==1]<-n1 n2[i==1]<-n2 n1[i+1]<-k1-alfa*n2[i] n2[i+1]<-k2-beta*n1[i] if(n1[i]==0){n1[i:t]==0} if(n2[i]==0){n2[i:t]==0} } plot(c(1:t), n1[1:t], type='b', xlim=range(c(1:t), c (1:t)), ylim=range(n1[1:t], n2[1:t]), xlab='tempo', ylab='tamanho populacional') points(c(1:t), n2[1:t], type='b', col="red") legend("topleft", c("Pop1", "Pop2"), cex=0.8, col=c ("black", "red"), pch=21:21, lty=1:1); }} Where n10: size population in time 0, n20: size population in time 0, k1: carrying capacity of the population 1, k2: carrying capacity of the population 2, alfa: competition coefficient of population 2 in population 1, beta: competition coefficient of population 1 in population 2, t: time. and when some population becomes 0 (ZERO), i want that population still 0 (ZERO) until the end of "t". i have tried to put " if(n1[i]==0){n1[i:t]==0} if(n2[i]==0){n2[i:t]==0}" after "n2[i+1]<-k2-beta*n1[i]" in the for function, but nothing happens. What may i do ? Thanks Bruno [[alternative HTML version deleted]] ------------------------------ Message: 16 Date: Sat, 24 Jul 2010 09:39:10 -0700 (PDT) From: Matt Stati <[109]mattstati at yahoo.com> To: [110]r-help at r-project.org Subject: [R] Book on R's Programming Language Message-ID: <[111]289386.7557.qm at web43507.mail.sp1.yahoo.com> Content-Type: text/plain Can someone please recommend to me a book on the programming language that R is based on? I'm looking for a foundational book that teaches the logic of the S language. It seems that knowing the underpinnings of the language can only make using R a bit easier. Any leads are greatly appreciated . . . Matt. [[alternative HTML version deleted]] ------------------------------ Message: 17 Date: Sat, 24 Jul 2010 13:39:53 -0300 From: Henrique Dallazuanna <[112]wwwhsd at gmail.com> To: aegea <[113]gcheer3 at gmail.com> Cc: [114]r-help at r-project.org Subject: Re: [R] how to calculate the product of every two elements in two vectors Message-ID: <AANLkTi=DCF=Jv9gzLQK8p=[115]XrwenTLja3ZTrwKE9USm4z at mail.gmail.com> Content-Type: text/plain Try this: c(as.matrix(B) %*% A) On Fri, Jul 23, 2010 at 12:11 PM, aegea <[116]gcheer3 at gmail.com> wrote: > > Thanks in advance! > > A=c(1, 2, 3) > B=c (9, 10, 11, 12) > > I want to get C=c(1*9, 1*10, 1*11, 1*12, ....., 3*9, 3*10, 3*11, 3*12)? > C is still a vector with 12 elements > Is there a way to do that? > -- > View this message in context: > [117]http://r.789695.n4.nabble.com/how-to-calculate-the-product-of-every-t wo-elements-in-two-vectors-tp2300299p2300299.html > Sent from the R help mailing list archive at Nabble.com. > > ______________________________________________ > [118]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > [119]http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. > -- Henrique Dallazuanna Curitiba-Paran???Brasil 25???25' 40" S 49???16' 22" O [[alternative HTML version deleted]] ------------------------------ Message: 18 Date: Sat, 24 Jul 2010 10:13:52 -0700 From: Joshua Wiley <[120]jwiley.psych at gmail.com> To: Matt Stati <[121]mattstati at yahoo.com> Cc: [122]r-help at r-project.org Subject: Re: [R] Book on R's Programming Language Message-ID: <AANLkTiki+9endGiR_T5pf9e8FofXbcyE5AL+-+[123]v0shNs at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 Hi Matt, [124]http://www.r-project.org/doc/bib/R-books.html Lists a variety of books, and seems to include most (i.e., my searches through google, amazon, and barnes and noble, didn't really turn up others) books dedicated to R. I have always been under the impression that Programming with Data (the Green Book) is a classic. [125]http://cran.r-project.org/manuals.html has the official manuals Similar questions have been asked several times on this list so you can also search for previous threads (e.g., [126]http://tolstoy.newcastle.edu.au/R/help/04/06/0063.html ) Best regards, Josh On Sat, Jul 24, 2010 at 9:39 AM, Matt Stati <[127]mattstati at yahoo.com> wrote: > Can someone please recommend to me a book on the programming language that R is based on? I'm looking for a foundational book that teaches the logic of the S language. It seems that knowing the underpinnings of the language can only make using R a bit easier. > > Any leads are greatly appreciated . . . > > Matt. > > > > > ? ? ? ?[[alternative HTML version deleted]] > > ______________________________________________ > [128]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide [129]http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. > -- Joshua Wiley Ph.D. Student, Health Psychology University of California, Los Angeles [130]http://www.joshuawiley.com/ ------------------------------ Message: 19 Date: Sat, 24 Jul 2010 13:15:34 -0400 From: Gabor Grothendieck <[131]ggrothendieck at gmail.com> To: aegea <[132]gcheer3 at gmail.com> Cc: [133]r-help at r-project.org Subject: Re: [R] how to calculate the product of every two elements in two vectors Message-ID: <AANLkTinRzLax+[134]dPbmg9YRPUcE2fQH_0Sqs2UJKFVJ3BS at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 On Fri, Jul 23, 2010 at 11:11 AM, aegea <[135]gcheer3 at gmail.com> wrote: > > Thanks in advance! > > A=c(1, 2, 3) > B=c (9, 10, 11, 12) > > I want to get C=c(1*9, 1*10, 1*11, 1*12, ....., 3*9, 3*10, 3*11, 3*12)? > C is still a vector with 12 elements > Is there a way to do that? Here are yet a few more. The first one is the only one so far that uses a single function and the last two are slight variations of ones already posted. kronecker(A, B) c(tcrossprod(B, A)) c(outer(B, A)) c(B %o% A) Here is a speed comparison. The fastest are as.matrix, %outer% and %o% . They are so close that random fluctuations might easily change their order and since %o% involves the least keystrokes that one might be a good overall choice. Although not among the fastest the kronecker solution is the simplest since it only involves a single function call so it might be preferred on that count. > A <- B <- 1:400 > out <- benchmark( + as.matrix = c(as.matrix(B) %*% A), + crossprod = c(tcrossprod(B, A)), + outer = c(outer(B, A)), + o = c(B %o% A), + kronecker = kronecker(A, B), + touter = as.vector(t(outer(A, B)))) > out[order(out$relative), ] test replications elapsed relative user.self sys.self user.child sys.child 1 as.matrix 100 0.92 1.000000 0.62 0.28 NA NA 3 outer 100 0.93 1.010870 0.59 0.35 NA NA 4 o 100 0.94 1.021739 0.66 0.28 NA NA 2 crossprod 100 1.11 1.206522 0.67 0.43 NA NA 5 kronecker 100 1.45 1.576087 1.25 0.21 NA NA 6 touter 100 1.84 2.000000 1.40 0.43 NA NA ------------------------------ Message: 20 Date: Sat, 24 Jul 2010 11:55:08 -0500 From: Joseph Magagnoli <[136]jcm331 at gmail.com> To: Matt Stati <[137]mattstati at yahoo.com>, rhelp <[138]r-help at r-project.org> Subject: Re: [R] Book on R's Programming Language Message-ID: <AANLkTi=+umQatto9rVeujEwjNyqOSiHOas+[139]eWg9Wcb80 at mail.gmail.com> Content-Type: text/plain Matt, you might want to check out programming with data by John Chambers. [140]http://www.amazon.com/Programming-Data-Guide-S-Language/dp/0387985034 /ref=sr_1_1?ie=UTF8&s=books&qid=1279990404&sr=8-1 Best Joe On Sat, Jul 24, 2010 at 11:39 AM, Matt Stati <[141]mattstati at yahoo.com> wrote: > Can someone please recommend to me a book on the programming language that > R is based on? I'm looking for a foundational book that teaches the logic of > the S language. It seems that knowing the underpinnings of the language can > only make using R a bit easier. > > Any leads are greatly appreciated . . . > > Matt. > > > > > [[alternative HTML version deleted]] > > ______________________________________________ > [142]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > [143]http://www.R-project.org/posting-guide.html<[144]http://www.r-project .org/posting-guide.html> > and provide commented, minimal, self-contained, reproducible code. > -- Joseph C. Magagnoli Doctoral Student Department of Political Science University of North Texas 1155 Union Circle #305340 Denton, Texas 76203-5017 Email: [145]jcm0250 at unt.edu [[alternative HTML version deleted]] ------------------------------ Message: 21 Date: Sat, 24 Jul 2010 12:55:18 -0600 From: Greg Snow <[146]Greg.Snow at imail.org> To: "Farley, Robert" <[147]FarleyR at metro.net>, "[148]r-help at r-project.org" <[149]r-help at r-project.org> Subject: Re: [R] Constrain density to 0 at 0? Message-ID: <[150]B37C0A15B8FB3C468B5BC7EBC7DA14CC633A53DF11 at LP-EXMBVS10.CO.IHC.COM> Content-Type: text/plain; charset="us-ascii" Look at the logspline package. This is a different approach to density estimation from the kernel densities used by 'density', but does allow you to set fixed boundaries. -- Gregory (Greg) L. Snow Ph.D. Statistical Data Center Intermountain Healthcare [151]greg.snow at imail.org 801.408.8111 > -----Original Message----- > From: [152]r-help-bounces at r-project.org [mailto:r-help-bounces at r- > project.org] On Behalf Of Farley, Robert > Sent: Monday, July 19, 2010 7:57 PM > To: [153]r-help at r-project.org > Subject: [R] Constrain density to 0 at 0? > > I'm plotting some trip length frequencies using the following code: > > plot( density(zTestData$Distance, weights=zTestData$Actual), > xlim=c(0, 10), > main="Test TLFD", > xlab="Distance", > col=6 ) > lines(density(zTestData$Distance, weights=zTestData$FlatWeight), col=2) > lines(density(zTestData$Distance, weights=zTestData$BrdWeight ), col=3) > > which works fine except the distances are all positive, but the > densities don't drop to 0 until around -2 or -3. > > Is there a way for me to "force" the density plot to 0 at 0? > > > > Thanks > > > > Robert Farley > Metro > 1 Gateway Plaza > Mail Stop 99-23-7 > Los Angeles, CA 90012-2952 > Voice: (213)922-2532 > Fax: (213)922-2868 > www.Metro.net > > > > [[alternative HTML version deleted]] > > ______________________________________________ > [154]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide [155]http://www.R-project.org/posting- > guide.html > and provide commented, minimal, self-contained, reproducible code. ------------------------------ Message: 22 Date: Sun, 25 Jul 2010 01:16:57 +0530 From: shabnam k <[156]shabnambioinfo at gmail.com> To: r-help <[157]r-help at r-project.org> Subject: [R] matest function for multiple factors Message-ID: <AANLkTi=mQ5bhntKQbfUapHpWNWpVLn+[158]Hzat_FD6jpZF6 at mail.gmail.com> Content-Type: text/plain Hi, I am using maanova package for analysis. In my dataset, two fixed factors time and treatment and sample as random factor is there. Am able to get madata object and fitmaanova object. But, am unable to do f-test with two factors, but i have done f-test seperately for two factors. fit.full.mix <- fitmaanova(madata, formula = ~Sample+Time+Treatment, random = ~Sample) ftest.all = *matest*(madata, fit.full.mix, test.method=c(1, 1), shuffle.method="sample", *term="Time+Treatment"*, n.perm= 100) Can u please suggest me, how to represent multiple factors in the above function simultaneously in term. [[alternative HTML version deleted]] ------------------------------ Message: 23 Date: Sat, 24 Jul 2010 13:55:59 -0600 From: Greg Snow <[159]Greg.Snow at imail.org> To: "[160]babyfoxlove1 at sina.com" <[161]babyfoxlove1 at sina.com>, "[162]r-help at r-project.org" <[163]r-help at r-project.org> Subject: Re: [R] How to deal with more than 6GB dataset using R? Message-ID: <[164]B37C0A15B8FB3C468B5BC7EBC7DA14CC633A53DF25 at LP-EXMBVS10.CO.IHC.COM> Content-Type: text/plain; charset="us-ascii" You may want to look at the biglm package as another way to regression models on very large data sets. -- Gregory (Greg) L. Snow Ph.D. Statistical Data Center Intermountain Healthcare [165]greg.snow at imail.org 801.408.8111 > -----Original Message----- > From: [166]r-help-bounces at r-project.org [mailto:r-help-bounces at r- > project.org] On Behalf Of [167]babyfoxlove1 at sina.com > Sent: Friday, July 23, 2010 10:10 AM > To: [168]r-help at r-project.org > Subject: [R] How to deal with more than 6GB dataset using R? > > Hi there, > > Sorry to bother those who are not interested in this problem. > > I'm dealing with a large data set, more than 6 GB file, and doing > regression test with those data. I was wondering are there any > efficient ways to read those data? Instead of just using read.table()? > BTW, I'm using a 64bit version desktop and a 64bit version R, and the > memory for the desktop is enough for me to use. > Thanks. > > > --Gin > > [[alternative HTML version deleted]] > > ______________________________________________ > [169]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide [170]http://www.R-project.org/posting- > guide.html > and provide commented, minimal, self-contained, reproducible code. ------------------------------ Message: 24 Date: Sat, 24 Jul 2010 17:03:44 -0400 From: "Abdi, Abdulhakim" <[171]AbdiA at si.edu> To: "[172]r-help at r-project.org" <[173]r-help at r-project.org> Subject: [R] Using R to fill ETM+ data gaps? Message-ID: <[174]97679C0A11332E48A01E0D463E8B3FF103AF0CE275 at SI-MSEV02.US.SINET.SI.EDU > Content-Type: text/plain; charset="us-ascii" Hi, I was wondering if anyone knows of a method (or if it's possible) to use R in interpolating Landsat ETM+ data gaps? Regards, Hakim Abdi _________________________________________ Abdulhakim Abdi, M.Sc. Conservation GIS/Remote Sensing Lab Smithsonian Conservation Biology Institute 1500 Remount Road Front Royal, VA 22630 phone: +1 540 635 6578 mobile: +1 747 224 7006 fax: +1 540 635 6506 (Attn: ABDI/GIS Lab) email: [175]abdia at si.edu [176]http://nationalzoo.si.edu/SCBI/ConservationGIS/ ------------------------------ Message: 25 Date: Sat, 24 Jul 2010 14:07:25 -0700 (PDT) From: Felipe Carrillo <[177]mazatlanmexico at yahoo.com> To: [178]r-help at stat.math.ethz.ch Subject: [R] How to generate a sequence of dates without hardcoding the year Message-ID: <[179]418059.32636.qm at web56602.mail.re3.yahoo.com> Content-Type: text/plain; charset=iso-8859-1 Hi: I have a dataframe named 'spring' and I am trying to add a new variable named 'IdDate' This line of code works fine: spring$idDate <- seq(as.Date("2008-07-01"), as.Date("2009-06-30"), by="week") But I don't want to hardcode the year because it will be used again the following year Is it possible to just generate dates with the month and day? I tried the code below: seq(as.Date("7-1", "%B%d"), as.Date("6-30", "%B%d"), by="week") and got this?error message: Error in seq.int(0, to - from, by) : 'to' must be finite Thanks for any pointers ? Felipe D. Carrillo Supervisory Fishery Biologist Department of the Interior US Fish & Wildlife Service California, USA ------------------------------ Message: 26 Date: Sat, 24 Jul 2010 18:09:45 -0300 From: Henrique Dallazuanna <[180]wwwhsd at gmail.com> To: Felipe Carrillo <[181]mazatlanmexico at yahoo.com> Cc: [182]r-help at stat.math.ethz.ch Subject: Re: [R] How to generate a sequence of dates without hardcoding the year Message-ID: <[183]AANLkTimd6DLagDiHKbFuF3hE5dES4d6DmC5T50Vn25N3 at mail.gmail.com> Content-Type: text/plain Try this: format(seq(as.Date("2008-07-01"), as.Date("2009-06-30"), by="week"), "%d/%m") On Sat, Jul 24, 2010 at 6:07 PM, Felipe Carrillo <[184]mazatlanmexico at yahoo.com>wrote: > Hi: > I have a dataframe named 'spring' and I am trying to add a new variable > named > 'IdDate' > This line of code works fine: > spring$idDate <- seq(as.Date("2008-07-01"), as.Date("2009-06-30"), by="week") > > But I don't want to hardcode the year because it will be used again the > following year > Is it possible to just generate dates with the month and day? > > I tried the code below: > seq(as.Date("7-1", "%B%d"), as.Date("6-30", "%B%d"), by="week") > > and got this error message: > Error in seq.int(0, to - from, by) : 'to' must be finite > Thanks for any pointers > > > Felipe D. Carrillo > Supervisory Fishery Biologist > Department of the Interior > US Fish & Wildlife Service > California, USA > > > > > ______________________________________________ > [185]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > [186]http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. > -- Henrique Dallazuanna Curitiba-Paran???Brasil 25???25' 40" S 49???16' 22" O [[alternative HTML version deleted]] ------------------------------ Message: 27 Date: Sat, 24 Jul 2010 15:54:51 -0500 (CDT) From: <[187]mpward at illinois.edu> To: "Joshua Wiley" <[188]jwiley.psych at gmail.com> Cc: [189]r-help at r-project.org Subject: Re: [R] Trouble retrieving the second largest value from each row of a data.frame Message-ID: <[190]20100724155451.CHG28413 at expms6.cites.uiuc.edu> Content-Type: text/plain; charset=iso-8859-1 THANKS, but I have one issue and one question. For some reason the "secondstrongest" value for row 3 and 6 are incorrect (they are the strongest) the remaining 10 are correct?? These data are being used to track radio-tagged birds, they are from automated radio telemetry receivers. I will applying the following formula diff <- ((strongest- secondstrongest)/100) bearingdiff <-30-(-0.0624*(diff**2))-(2.8346*diff) Then the bearing diff is added to strongestantenna (value0 = 0degrees) if the secondstrongestatenna is greater (eg value0 and value60), or if the secondstrongestantenna is smaller than the strongestantenna, then the bearingdiff is substracted from the strongestantenna. The only exception is that if value0 (0degrees) is strongest and value300(360degrees) is the secondstrongestantenna then the bearing is 360-bearingdiff. Also the strongestantenna and secondstrongestantenna have to be next to each other (e.g. value0 with value60, value240 with value300, value0 with value300) or the results should be NA. I have been trying to use a series of if, else statements to produce these bearing, but all I am producing is errors. Any suggestion would be appreciated. Again THANKS for you efforts. Mike ---- Original message ---- >Date: Fri, 23 Jul 2010 23:01:56 -0700 >From: Joshua Wiley <[191]jwiley.psych at gmail.com> >Subject: Re: [R] Trouble retrieving the second largest value from each row of a data.frame >To: [192]mpward at illinois.edu >Cc: [193]r-help at r-project.org > >Hi, > >Here is a little function that will do what you want and return a nice output: > >#Function To calculate top two values and return >my.finder <- function(mydata) { > my.fun <- function(data) { > strongest <- which.max(data) > secondstrongest <- which.max(data[-strongest]) > strongestantenna <- names(data)[strongest] > secondstrongantenna <- names(data[-strongest])[secondstrongest] > value <- matrix(c(data[strongest], data[secondstrongest], > strongestantenna, secondstrongantenna), ncol =4) > return(value) > } > dat <- apply(mydata, 1, my.fun) > dat <- t(dat) > dat <- as.data.frame(dat, stringsAsFactors = FALSE) > colnames(dat) <- c("strongest", "secondstrongest", > "strongestantenna", "secondstrongantenna") > dat[ , "strongest"] <- as.numeric(dat[ , "strongest"]) > dat[ , "secondstrongest"] <- as.numeric(dat[ , "secondstrongest"]) > return(dat) >} > > >#Using your example data: > >yourdata <- structure(list(value0 = c(-13007L, -12838L, -12880L, -12805L, >-12834L, -11068L, -12807L, -12770L, -12988L, -11779L), value60 c(-11707L, >-13210L, -11778L, -11653L, -13527L, -11698L, -14068L, -11665L, >-11736L, -12873L), value120 = c(-11072L, -11176L, -11113L, -11071L, >-11067L, -12430L, -11092L, -11061L, -11137L, -12973L), value180 c(-12471L, >-11799L, -12439L, -12385L, -11638L, -12430L, -11709L, -12373L, >-12570L, -12537L), value240 = c(-12838L, -13210L, -13089L, -11561L, >-13527L, -12430L, -11607L, -11426L, -13467L, -12973L), value300 c(-13357L, >-13845L, -13880L, -13317L, -13873L, -12814L, -13025L, -12805L, >-13739L, -11146L)), .Names = c("value0", "value60", "value120", >"value180", "value240", "value300"), class = "data.frame", row.names c("1", >"2", "3", "4", "5", "6", "7", "8", "9", "10")) > >my.finder(yourdata) #and what you want is in a nicely labeled data frame > >#A potential problem is that it is not very efficient > >#Here is a test using a matrix of 100, 000 rows >#sampled from the same range as your data >#with the same number of columns > >data.test <- matrix( > sample(seq(min(yourdata), max(yourdata)), size = 500000, replace TRUE), > ncol = 5) > >system.time(my.finder(data.test)) > >#On my system I get > >> system.time(my.finder(data.test)) > user system elapsed > 2.89 0.00 2.89 > >Hope that helps, > >Josh > > > >On Fri, Jul 23, 2010 at 6:20 PM, <[194]mpward at illinois.edu> wrote: >> I have a data frame with a couple million lines and want to retrieve the largest and second largest values in each row, along with the label of the column these values are in. For example >> >> row 1 >> strongest=-11072 >> secondstrongest=-11707 >> strongestantenna=value120 >> secondstrongantenna=value60 >> >> Below is the code I am using and a truncated data.frame. ?Retrieving the largest value was easy, but I have been getting errors every way I have tried to retrieve the second largest value. ?I have not even tried to retrieve the labels for the value yet. >> >> Any help would be appreciated >> Mike >> >> >>> data<-data.frame(value0, value60, value120, value180, value240, value300) >>> data >> ? value0 value60 value120 value180 value240 value300 >> 1 ?-13007 ?-11707 ? -11072 ? -12471 ? -12838 ? -13357 >> 2 ?-12838 ?-13210 ? -11176 ? -11799 ? -13210 ? -13845 >> 3 ?-12880 ?-11778 ? -11113 ? -12439 ? -13089 ? -13880 >> 4 ?-12805 ?-11653 ? -11071 ? -12385 ? -11561 ? -13317 >> 5 ?-12834 ?-13527 ? -11067 ? -11638 ? -13527 ? -13873 >> 6 ?-11068 ?-11698 ? -12430 ? -12430 ? -12430 ? -12814 >> 7 ?-12807 ?-14068 ? -11092 ? -11709 ? -11607 ? -13025 >> 8 ?-12770 ?-11665 ? -11061 ? -12373 ? -11426 ? -12805 >> 9 ?-12988 ?-11736 ? -11137 ? -12570 ? -13467 ? -13739 >> 10 -11779 ?-12873 ? -12973 ? -12537 ? -12973 ? -11146 >>> #largest value in the row >>> strongest<-apply(data, 1, max) >>> >>> >>> #second largest value in the row >>> n<-function(data)(1/(min(1/(data[1, ]-max(data[1, ]))))+ (max(data[1, ]))) >>> secondstrongest<-apply(data, 1, n) >> Error in data[1, ] : incorrect number of dimensions >>> >> >> ______________________________________________ >> [195]R-help at r-project.org mailing list >> https://stat.ethz.ch/mailman/listinfo/r-help >> PLEASE do read the posting guide [196]http://www.R-project.org/posting-guide.html >> and provide commented, minimal, self-contained, reproducible code. >> > > > >-- >Joshua Wiley >Ph.D. Student, Health Psychology >University of California, Los Angeles >[197]http://www.joshuawiley.com/ ------------------------------ Message: 28 Date: Sat, 24 Jul 2010 22:21:21 +0100 From: Paul Smith <[198]phhs80 at gmail.com> To: [199]r-help at r-project.org Subject: Re: [R] (no subject) Message-ID: <AANLkTi=qKpbPZutDNr9n8af1jGFsLfTqT1C7Nm8Ko+[200]oH at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 2010/7/23 w ? <[201]hw_joyce_cn at hotmail.com>: > I use the constrOptim to maximize a function with four constriants but the answer does not leave from the starting value and there is only one outer iteration. The function is defined as follows: > tm<-function(p){ > p1<-p[1]; p2<-p[2]; p3<-p[3]; > p4<-1-p1-p2-p3; > p1*p2*p3*p4} > > ##the constraints are p1>=0; p2>=0; p3>=0 and p4>=0 i.e. p1+p2+p3<=1 > start<-c(0.9999991, 0.0000001, 0.0000001) > dev<-rbind(diag(3), -diag(3), rep(-1, 3)) > bvec<-c(rep(0, 3), rep(-1, 4)) > constrOptim(start, tm, NULL, ui=dev, ci=bvec, control=list(maxit=10000)) > > Am i missing something obviously that cause the problem or there is some bugs in constrOptim. Could you please help me out Wenwen, I believe that the reason why constrOptim behaves as described is related to the fact that (p1, p2, p3) = (1, 0, 0) is a stationary point and you use it as a starting point. Try a different starting point. If the objective is to maximize, then you should use the following command: constrOptim(start, tm, NULL, ui=dev, ci=bvec, control=list(maxit=10000, fnscale=-1)) (Notice fnscale=-1.) Finally, whenever you ask something on this list, please use a meaningful title for your message, as it will dramatically increase the chances of you getting an answer. Good luck, Paul ------------------------------ Message: 29 Date: Sat, 24 Jul 2010 19:02:57 -0400 From: jim holtman <[202]jholtman at gmail.com> To: Felipe Carrillo <[203]mazatlanmexico at yahoo.com> Cc: [204]r-help at stat.math.ethz.ch Subject: Re: [R] How to generate a sequence of dates without hardcoding the year Message-ID: <AANLkTikBNcm2=EzKaWMGNez-b53QhYYe_9CuK4LRN=[205]3P at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 Is this what you want if you want to assume that the date without a year is this year: > seq(as.Date("7-1", "%m-%d"), by="week", length=52) [1] "2010-07-01" "2010-07-08" "2010-07-15" "2010-07-22" "2010-07-29" "2010-08-05" "2010-08-12" "2010-08-19" [9] "2010-08-26" "2010-09-02" "2010-09-09" "2010-09-16" "2010-09-23" "2010-09-30" "2010-10-07" "2010-10-14" [17] "2010-10-21" "2010-10-28" "2010-11-04" "2010-11-11" "2010-11-18" "2010-11-25" "2010-12-02" "2010-12-09" [25] "2010-12-16" "2010-12-23" "2010-12-30" "2011-01-06" "2011-01-13" "2011-01-20" "2011-01-27" "2011-02-03" [33] "2011-02-10" "2011-02-17" "2011-02-24" "2011-03-03" "2011-03-10" "2011-03-17" "2011-03-24" "2011-03-31" [41] "2011-04-07" "2011-04-14" "2011-04-21" "2011-04-28" "2011-05-05" "2011-05-12" "2011-05-19" "2011-05-26" [49] "2011-06-02" "2011-06-09" "2011-06-16" "2011-06-23" > On Sat, Jul 24, 2010 at 5:07 PM, Felipe Carrillo <[206]mazatlanmexico at yahoo.com> wrote: > Hi: > I have a dataframe named 'spring' and I am trying to add a new variable named > 'IdDate' > This line of code works fine: > spring$idDate <- seq(as.Date("2008-07-01"), as.Date("2009-06-30"), by="week") > > But I don't want to hardcode the year because it will be used again the > following year > Is it possible to just generate dates with the month and day? > > I tried the code below: > seq(as.Date("7-1", "%B%d"), as.Date("6-30", "%B%d"), by="week") > > and got this?error message: > Error in seq.int(0, to - from, by) : 'to' must be finite > Thanks for any pointers > > > Felipe D. Carrillo > Supervisory Fishery Biologist > Department of the Interior > US Fish & Wildlife Service > California, USA > > > > > ______________________________________________ > [207]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide [208]http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. > -- Jim Holtman Cincinnati, OH +1 513 646 9390 What is the problem that you are trying to solve? ------------------------------ Message: 30 Date: Sat, 24 Jul 2010 16:27:06 -0700 (PDT) From: zachmohr <[209]zachmohr at gmail.com> To: [210]r-help at r-project.org Subject: Re: [R] glm - prediction of a factor with several levels Message-ID: <[211]AANLkTinzngqO3S_p-E25N3TTWEdiQBBS4CvzErgywj1p at mail.gmail.com> Content-Type: text/plain As far as I know, glm only works with dichotomous or count data. polr in the MASS package works and so does lrm {Design} for ordinal dependent variables. I would assume that the model produced by glm is a dichotomous version of your model but not sure. Only one intercept would be given because if you used the log link then it would have produced a dichotomous model instead of an ordered logistic regression. My suggestion is to use polr or lrm. On Fri, Jul 23, 2010 at 6:15 PM, blackscorpio [via R] < ml-node+[212]2300793-1751019155-246278 at n4.nabble.com<ml-node%[213]2B230079 3-1751019155-246278 at n4.nabble.com> > wrote: > Dear community, > I'm currently attempting to predict the occurence of an event (factor) > having more than 2 levels with several continuous predictors. The model > being ordinal, I was waiting the glm function to return several intercepts, > which is not the case when looking to my results (I only have one > intercept). I finally managed to perform an ordinal polytomous logisitc > regression with the polr function, which gives several intercepts. > But does anyone know what was the model performed with glm and why only one > intercept was given ? > Thanks a lot for your help ! > > > ------------------------------ > View message @ > [214]http://r.789695.n4.nabble.com/glm-prediction-of-a-factor-with-several -levels-tp2300793p2300793.html > To unsubscribe from R, click here< (link removed) ==>. > > > -- Department of Public Administration University of Kansas 4060 Wesco Hall Office W Lawrence KS 66045-3177 Phone: (785) 813-1384 Email: [215]zachmohr at gmail.com -- View this message in context: [216]http://r.789695.n4.nabble.com/glm-prediction-of-a-factor-with-several -levels-tp2300793p2301324.html Sent from the R help mailing list archive at Nabble.com. [[alternative HTML version deleted]] ------------------------------ Message: 31 Date: Sat, 24 Jul 2010 20:09:32 -0400 From: David Winsemius <[217]dwinsemius at comcast.net> To: <[218]mpward at illinois.edu> Cc: [219]r-help at r-project.org Subject: Re: [R] Trouble retrieving the second largest value from each row of a data.frame Message-ID: <[220]52EA484F-C066-4ACC-B5BC-1A3A20876D9E at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Jul 24, 2010, at 4:54 PM, <[221]mpward at illinois.edu> wrote: > THANKS, but I have one issue and one question. > > For some reason the "secondstrongest" value for row 3 and 6 are > incorrect (they are the strongest) the remaining 10 are correct?? In my run of Wiley's code I instead get identical values for rows 2, 5, 6. Holtman's and my solutions did not suffer from that defect, although mine suffered from my misreading of your request, thinking that you wanted the top 3. The fix is trivial > > These data are being used to track radio-tagged birds, they are from > automated radio telemetry receivers. I will applying the following > formula > > diff <- ((strongest- secondstrongest)/100) > bearingdiff <-30-(-0.0624*(diff**2))-(2.8346*diff) vals <- c("value0", "value60", "value120", "value180", "value240", "value300") value.str2 <- (match(yourdata$secondstrongestantenna, vals)-1)*60 value.str1 <- (match(yourdata$strongestantenna, vals)-1)*60 change.ind <- abs(match(yourdata, vals) - which(match(yourdata, vals) ) > > A) Then the bearing diff is added to strongestantenna (value0 > 0degrees) if the secondstrongestatenna is greater (eg value0 and > value60), > B) or if the secondstrongestantenna is smaller than the > strongestantenna, > then the bearingdiff is substracted from the strongestantenna. > > C) The only exception is that if value0 (0degrees) is strongest and > value300(360degrees) is the secondstrongestantenna then the bearing > is 360-bearingdiff. > D) Also the strongestantenna and secondstrongestantenna have to be > next to each other (e.g. value0 with value60, value240 with > value300, value0 with value300) or the results should be NA. After setting finalbearing with A, B, and C then: yourdata$finalbearing <- with(yourdata, ifelse ( change.ind <5 & change.ind > 1 , NA, finalbearing) ) > I have been trying to use a series of if, else statements to produce > these bearing, but all I am producing is errors. Any suggestion > would be appreciated. > > Again THANKS for you efforts. > > Mike > > ---- Original message ---- >> Date: Fri, 23 Jul 2010 23:01:56 -0700 >> From: Joshua Wiley <[222]jwiley.psych at gmail.com> >> Subject: Re: [R] Trouble retrieving the second largest value from >> each row of a data.frame >> To: [223]mpward at illinois.edu >> Cc: [224]r-help at r-project.org >> >> Hi, >> >> Here is a little function that will do what you want and return a >> nice output: >> >> #Function To calculate top two values and return >> my.finder <- function(mydata) { >> my.fun <- function(data) { >> strongest <- which.max(data) >> secondstrongest <- which.max(data[-strongest]) >> strongestantenna <- names(data)[strongest] >> secondstrongantenna <- names(data[-strongest])[secondstrongest] >> value <- matrix(c(data[strongest], data[secondstrongest], >> strongestantenna, secondstrongantenna), ncol =4) >> return(value) >> } >> dat <- apply(mydata, 1, my.fun) >> dat <- t(dat) >> dat <- as.data.frame(dat, stringsAsFactors = FALSE) >> colnames(dat) <- c("strongest", "secondstrongest", >> "strongestantenna", "secondstrongantenna") >> dat[ , "strongest"] <- as.numeric(dat[ , "strongest"]) >> dat[ , "secondstrongest"] <- as.numeric(dat[ , "secondstrongest"]) >> return(dat) >> } >> >> >> #Using your example data: >> >> yourdata <- structure(list(value0 = c(-13007L, -12838L, -12880L, >> -12805L, >> -12834L, -11068L, -12807L, -12770L, -12988L, -11779L), value60 >> c(-11707L, >> -13210L, -11778L, -11653L, -13527L, -11698L, -14068L, -11665L, >> -11736L, -12873L), value120 = c(-11072L, -11176L, -11113L, -11071L, >> -11067L, -12430L, -11092L, -11061L, -11137L, -12973L), value180 >> c(-12471L, >> -11799L, -12439L, -12385L, -11638L, -12430L, -11709L, -12373L, >> -12570L, -12537L), value240 = c(-12838L, -13210L, -13089L, -11561L, >> -13527L, -12430L, -11607L, -11426L, -13467L, -12973L), value300 >> c(-13357L, >> -13845L, -13880L, -13317L, -13873L, -12814L, -13025L, -12805L, >> -13739L, -11146L)), .Names = c("value0", "value60", "value120", >> "value180", "value240", "value300"), class = "data.frame", >> row.names = c("1", >> "2", "3", "4", "5", "6", "7", "8", "9", "10")) >> >> my.finder(yourdata) #and what you want is in a nicely labeled data >> frame >> >> #A potential problem is that it is not very efficient >> >> #Here is a test using a matrix of 100, 000 rows >> #sampled from the same range as your data >> #with the same number of columns >> >> data.test <- matrix( >> sample(seq(min(yourdata), max(yourdata)), size = 500000, replace >> TRUE), >> ncol = 5) >> >> system.time(my.finder(data.test)) >> >> #On my system I get >> >>> system.time(my.finder(data.test)) >> user system elapsed >> 2.89 0.00 2.89 >> >> Hope that helps, >> >> Josh >> >> >> >> On Fri, Jul 23, 2010 at 6:20 PM, <[225]mpward at illinois.edu> wrote: >>> I have a data frame with a couple million lines and want to >>> retrieve the largest and second largest values in each row, along >>> with the label of the column these values are in. For example >>> >>> row 1 >>> strongest=-11072 >>> secondstrongest=-11707 >>> strongestantenna=value120 >>> secondstrongantenna=value60 >>> >>> Below is the code I am using and a truncated data.frame. >>> Retrieving the largest value was easy, but I have been getting >>> errors every way I have tried to retrieve the second largest >>> value. I have not even tried to retrieve the labels for the value >>> yet. >>> >>> Any help would be appreciated >>> Mike >>> >>> >>>> data<- >>>> data.frame(value0, value60, value120, value180, value240, value300) >>>> data >>> value0 value60 value120 value180 value240 value300 >>> 1 -13007 -11707 -11072 -12471 -12838 -13357 >>> 2 -12838 -13210 -11176 -11799 -13210 -13845 >>> 3 -12880 -11778 -11113 -12439 -13089 -13880 >>> 4 -12805 -11653 -11071 -12385 -11561 -13317 >>> 5 -12834 -13527 -11067 -11638 -13527 -13873 >>> 6 -11068 -11698 -12430 -12430 -12430 -12814 >>> 7 -12807 -14068 -11092 -11709 -11607 -13025 >>> 8 -12770 -11665 -11061 -12373 -11426 -12805 >>> 9 -12988 -11736 -11137 -12570 -13467 -13739 >>> 10 -11779 -12873 -12973 -12537 -12973 -11146 >>>> #largest value in the row >>>> strongest<-apply(data, 1, max) >>>> >>>> >>>> #second largest value in the row >>>> n<-function(data)(1/(min(1/(data[1, ]-max(data[1, ]))))+ >>>> (max(data[1, ]))) >>>> secondstrongest<-apply(data, 1, n) >>> Error in data[1, ] : incorrect number of dimensions >>>> >>> >>> ______________________________________________ >>> [226]R-help at r-project.org mailing list >>> https://stat.ethz.ch/mailman/listinfo/r-help >>> PLEASE do read the posting guide [227]http://www.R-project.org/posting-guide.html >>> and provide commented, minimal, self-contained, reproducible code. >>> >> >> >> >> -- >> Joshua Wiley >> Ph.D. Student, Health Psychology >> University of California, Los Angeles >> [228]http://www.joshuawiley.com/ > > ______________________________________________ > [229]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide [230]http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. ------------------------------ Message: 32 Date: Sat, 24 Jul 2010 17:57:10 -0700 From: Joshua Wiley <[231]jwiley.psych at gmail.com> To: [232]mpward at illinois.edu Cc: [233]r-help at r-project.org Subject: Re: [R] Trouble retrieving the second largest value from each row of a data.frame Message-ID: <AANLkTikG3SgD+af50n2dsHMA3+[234]fa9gCQLuYBAaMS_P8w at mail.gmail.com> Content-Type: text/plain; charset=UTF-8 On Sat, Jul 24, 2010 at 5:09 PM, David Winsemius <[235]dwinsemius at comcast.net> wrote: > > On Jul 24, 2010, at 4:54 PM, <[236]mpward at illinois.edu> wrote: > >> THANKS, but I have one issue and one question. >> >> For some reason the "secondstrongest" value for row 3 and 6 are incorrect >> (they are the strongest) the remaining 10 are correct?? > > In my run of Wiley's code I instead get identical values for rows 2, 5, 6. Yes, my apologies; I neglected a [-strongest] when extracting the second highest value. I included a corrected form below; however, Winsemius' code is cleaner, not to mention easier to generalize, so I see no reason not to use that option. You might consider using a different object name than 'diff' since it is also the name of a function. Josh ####### my.finder <- function(mydata) { my.fun <- function(data) { strongest <- which.max(data) secondstrongest <- which.max(data[-strongest]) strongestantenna <- names(data)[strongest] secondstrongantenna <- names(data[-strongest])[secondstrongest] value <- matrix(c(data[strongest], data[-strongest][secondstrongest], strongestantenna, secondstrongantenna), ncol =4) return(value) } dat <- apply(mydata, 1, my.fun) dat <- t(dat) dat <- as.data.frame(dat, stringsAsFactors = FALSE) colnames(dat) <- c("strongest", "secondstrongest", "strongestantenna", "secondstrongantenna") dat[ , "strongest"] <- as.numeric(dat[ , "strongest"]) dat[ , "secondstrongest"] <- as.numeric(dat[ , "secondstrongest"]) return(dat) } > Holtman's and my solutions did not suffer from that defect, although mine > suffered from my misreading of your request, thinking that you wanted the > top 3. The fix is trivial >> >> These data are being used to track radio-tagged birds, they are from >> automated radio telemetry receivers. ?I will applying the following formula >> >> ?diff <- ((strongest- secondstrongest)/100) >> ?bearingdiff <-30-(-0.0624*(diff**2))-(2.8346*diff) > > vals <- c("value0", "value60", "value120", "value180", "value240", > "value300") > value.str2 <- (match(yourdata$secondstrongestantenna, vals)-1)*60 > value.str1 <- (match(yourdata$strongestantenna, vals)-1)*60 > change.ind <- abs(match(yourdata, vals) - which(match(yourdata, vals) ) > >> >> A) Then the bearing diff is added to strongestantenna (value0 0degrees) >> if the secondstrongestatenna is greater (eg value0 and value60), > >> B) or if the secondstrongestantenna is smaller than the strongestantenna, >> then the bearingdiff is substracted from the strongestantenna. > >> >> C) The only exception is that if value0 (0degrees) is strongest and >> value300(360degrees) is the secondstrongestantenna then the bearing is >> 360-bearingdiff. > > >> D) Also the strongestantenna and secondstrongestantenna have to be next to >> each other (e.g. value0 with value60, value240 with value300, value0 with >> value300) or the results should be NA. > > After setting finalbearing with A, B, and C then: > yourdata$finalbearing <- with(yourdata, ifelse ( > ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ?change.ind <5 & change.ind > 1 , > ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? NA, finalbearing) ) > >> I have been trying to use a series of if, else statements to produce these >> bearing, but all I am producing is errors. Any suggestion would be >> appreciated. > > >> >> Again THANKS for you efforts. >> >> Mike >> >> ---- Original message ---- >>> >>> Date: Fri, 23 Jul 2010 23:01:56 -0700 >>> From: Joshua Wiley <[237]jwiley.psych at gmail.com> >>> Subject: Re: [R] Trouble retrieving the second largest value from each >>> row of ?a data.frame >>> To: [238]mpward at illinois.edu >>> Cc: [239]r-help at r-project.org >>> >>> Hi, >>> >>> Here is a little function that will do what you want and return a nice >>> output: >>> >>> #Function To calculate top two values and return >>> my.finder <- function(mydata) { >>> my.fun <- function(data) { >>> ?strongest <- which.max(data) >>> ?secondstrongest <- which.max(data[-strongest]) >>> ?strongestantenna <- names(data)[strongest] >>> ?secondstrongantenna <- names(data[-strongest])[secondstrongest] >>> ?value <- matrix(c(data[strongest], data[secondstrongest], >>> ? ? ? ? ? ? ? ? ? ?strongestantenna, secondstrongantenna), ncol =4) >>> ?return(value) >>> } >>> dat <- apply(mydata, 1, my.fun) >>> dat <- t(dat) >>> dat <- as.data.frame(dat, stringsAsFactors = FALSE) >>> colnames(dat) <- c("strongest", "secondstrongest", >>> ? ? ? ? ? ? ? ? ? "strongestantenna", "secondstrongantenna") >>> dat[ , "strongest"] <- as.numeric(dat[ , "strongest"]) >>> dat[ , "secondstrongest"] <- as.numeric(dat[ , "secondstrongest"]) >>> return(dat) >>> } >>> >>> >>> #Using your example data: >>> >>> yourdata <- structure(list(value0 = c(-13007L, -12838L, -12880L, -12805L, >>> -12834L, -11068L, -12807L, -12770L, -12988L, -11779L), value60 >>> c(-11707L, >>> -13210L, -11778L, -11653L, -13527L, -11698L, -14068L, -11665L, >>> -11736L, -12873L), value120 = c(-11072L, -11176L, -11113L, -11071L, >>> -11067L, -12430L, -11092L, -11061L, -11137L, -12973L), value180 >>> c(-12471L, >>> -11799L, -12439L, -12385L, -11638L, -12430L, -11709L, -12373L, >>> -12570L, -12537L), value240 = c(-12838L, -13210L, -13089L, -11561L, >>> -13527L, -12430L, -11607L, -11426L, -13467L, -12973L), value300 >>> c(-13357L, >>> -13845L, -13880L, -13317L, -13873L, -12814L, -13025L, -12805L, >>> -13739L, -11146L)), .Names = c("value0", "value60", "value120", >>> "value180", "value240", "value300"), class = "data.frame", row.names >>> c("1", >>> "2", "3", "4", "5", "6", "7", "8", "9", "10")) >>> >>> my.finder(yourdata) #and what you want is in a nicely labeled data frame >>> >>> #A potential problem is that it is not very efficient >>> >>> #Here is a test using a matrix of 100, 000 rows >>> #sampled from the same range as your data >>> #with the same number of columns >>> >>> data.test <- matrix( >>> sample(seq(min(yourdata), max(yourdata)), size = 500000, replace TRUE), >>> ncol = 5) >>> >>> system.time(my.finder(data.test)) >>> >>> #On my system I get >>> >>>> system.time(my.finder(data.test)) >>> >>> ?user ?system elapsed >>> ?2.89 ? ?0.00 ? ?2.89 >>> >>> Hope that helps, >>> >>> Josh >>> >>> >>> >>> On Fri, Jul 23, 2010 at 6:20 PM, ?<[240]mpward at illinois.edu> wrote: >>>> >>>> I have a data frame with a couple million lines and want to retrieve the >>>> largest and second largest values in each row, along with the label of the >>>> column these values are in. For example >>>> >>>> row 1 >>>> strongest=-11072 >>>> secondstrongest=-11707 >>>> strongestantenna=value120 >>>> secondstrongantenna=value60 >>>> >>>> Below is the code I am using and a truncated data.frame. ?Retrieving the >>>> largest value was easy, but I have been getting errors every way I have >>>> tried to retrieve the second largest value. ?I have not even tried to >>>> retrieve the labels for the value yet. >>>> >>>> Any help would be appreciated >>>> Mike >>>> >>>> >>>>> data<-data.frame(value0, value60, value120, value180, value240, value300) >>>>> data >>>> >>>> ?value0 value60 value120 value180 value240 value300 >>>> 1 ?-13007 ?-11707 ? -11072 ? -12471 ? -12838 ? -13357 >>>> 2 ?-12838 ?-13210 ? -11176 ? -11799 ? -13210 ? -13845 >>>> 3 ?-12880 ?-11778 ? -11113 ? -12439 ? -13089 ? -13880 >>>> 4 ?-12805 ?-11653 ? -11071 ? -12385 ? -11561 ? -13317 >>>> 5 ?-12834 ?-13527 ? -11067 ? -11638 ? -13527 ? -13873 >>>> 6 ?-11068 ?-11698 ? -12430 ? -12430 ? -12430 ? -12814 >>>> 7 ?-12807 ?-14068 ? -11092 ? -11709 ? -11607 ? -13025 >>>> 8 ?-12770 ?-11665 ? -11061 ? -12373 ? -11426 ? -12805 >>>> 9 ?-12988 ?-11736 ? -11137 ? -12570 ? -13467 ? -13739 >>>> 10 -11779 ?-12873 ? -12973 ? -12537 ? -12973 ? -11146 >>>>> >>>>> #largest value in the row >>>>> strongest<-apply(data, 1, max) >>>>> >>>>> >>>>> #second largest value in the row >>>>> n<-function(data)(1/(min(1/(data[1, ]-max(data[1, ]))))+ (max(data[1, ]))) >>>>> secondstrongest<-apply(data, 1, n) >>>> >>>> Error in data[1, ] : incorrect number of dimensions >>>>> >>>> >>>> ______________________________________________ >>>> [241]R-help at r-project.org mailing list >>>> https://stat.ethz.ch/mailman/listinfo/r-help >>>> PLEASE do read the posting guide >>>> [242]http://www.R-project.org/posting-guide.html >>>> and provide commented, minimal, self-contained, reproducible code. >>>> >>> >>> >>> >>> -- >>> Joshua Wiley >>> Ph.D. Student, Health Psychology >>> University of California, Los Angeles >>> [243]http://www.joshuawiley.com/ >> >> ______________________________________________ >> [244]R-help at r-project.org mailing list >> https://stat.ethz.ch/mailman/listinfo/r-help >> PLEASE do read the posting guide >> [245]http://www.R-project.org/posting-guide.html >> and provide commented, minimal, self-contained, reproducible code. > > -- Joshua Wiley Ph.D. Student, Health Psychology University of California, Los Angeles [246]http://www.joshuawiley.com/ ------------------------------ Message: 33 Date: Sat, 24 Jul 2010 21:27:17 -0400 From: David Winsemius <[247]dwinsemius at comcast.net> To: "[248]r-help at r-project.org list" <[249]r-help at r-project.org> Cc: [250]mpward at illinois.edu Subject: Re: [R] Trouble retrieving the second largest value from each row of a data.frame Message-ID: <[251]2B19FDC3-4358-4731-87C7-89399E5DD75E at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Jul 24, 2010, at 8:09 PM, David Winsemius wrote: > > On Jul 24, 2010, at 4:54 PM, <[252]mpward at illinois.edu> wrote: > >> THANKS, but I have one issue and one question. >> >> For some reason the "secondstrongest" value for row 3 and 6 are >> incorrect (they are the strongest) the remaining 10 are correct?? > > In my run of Wiley's code I instead get identical values for rows > 2, 5, 6. Holtman's and my solutions did not suffer from that defect, > although mine suffered from my misreading of your request, thinking > that you wanted the top 3. The fix is trivial >> >> These data are being used to track radio-tagged birds, they are >> from automated radio telemetry receivers. I will applying the >> following formula >> >> diff <- ((strongest- secondstrongest)/100) >> bearingdiff <-30-(-0.0624*(diff**2))-(2.8346*diff) > > vals <- c("value0", "value60", "value120", "value180", "value240", > "value300") > value.str2 <- (match(yourdata$secondstrongestantenna, vals)-1)*60 > value.str1 <- (match(yourdata$strongestantenna, vals)-1)*60 > change.ind <- abs(match(yourdata, vals) - which(match(yourdata, > vals) ) OOOPs should have been change.ind <- abs(match(yourdata, vals) - match(yourdata, vals) ) > >> >> A) Then the bearing diff is added to strongestantenna (value0 >> 0degrees) if the secondstrongestatenna is greater (eg value0 and >> value60), > >> B) or if the secondstrongestantenna is smaller than the >> strongestantenna, >> then the bearingdiff is substracted from the strongestantenna. yourdata$finalbearing <- with(yourdata, ifelse (value.str2>value.str1, bearingdiff+value.str1, value.str1-bearingdiff) ) > >> >> C) The only exception is that if value0 (0degrees) is strongest and >> value300(360degrees) is the secondstrongestantenna then the bearing >> is 360-bearingdiff. > yourdata$finalbearing <- with(yourdata, ifelse (strongestantenna = "value0" & secondstrongantenna == "value300", 360- bearingdiff, finalbearing) ); >> D) Also the strongestantenna and secondstrongestantenna have to be >> next to each other (e.g. value0 with value60, value240 with >> value300, value0 with value300) or the results should be NA. > > After setting finalbearing with A, B, and C then: > yourdata$finalbearing <- with(yourdata, ifelse ( > change.ind <5 & change.ind > 1 , > NA, finalbearing) ) > yourdata strongest secondstrongest strongestantenna secondstrongantenna finalbearing 1 -11072 -11707 value120 value60 -11086.52 2 -11176 -11799 value120 value180 -11190.76 3 -11113 -11778 value120 value60 -11126.91 4 -11071 -11561 value120 value240 NA 5 -11067 -11638 value120 value180 -11082.85 6 -11068 -11698 value0 value60 -11082.62 7 -11092 -11607 value120 value240 NA 8 -11061 -11426 value120 value240 NA 9 -11137 -11736 value120 value60 -11152.26 10 -11146 -11779 value300 value0 -11160.56 > >> I have been trying to use a series of if, else statements to produce >> these bearing, ifelse is the correct construct for processing vectors -- David. >> but all I am producing is errors. Any suggestion would be >> appreciated. > > >> >> Again THANKS for you efforts. >> >> Mike >> >> ---- Original message ---- >>> Date: Fri, 23 Jul 2010 23:01:56 -0700 >>> From: Joshua Wiley <[253]jwiley.psych at gmail.com> >>> Subject: Re: [R] Trouble retrieving the second largest value from >>> each row of a data.frame >>> To: [254]mpward at illinois.edu >>> Cc: [255]r-help at r-project.org >>> >>> Hi, >>> >>> Here is a little function that will do what you want and return a >>> nice output: >>> >>> #Function To calculate top two values and return >>> my.finder <- function(mydata) { >>> my.fun <- function(data) { >>> strongest <- which.max(data) >>> secondstrongest <- which.max(data[-strongest]) >>> strongestantenna <- names(data)[strongest] >>> secondstrongantenna <- names(data[-strongest])[secondstrongest] >>> value <- matrix(c(data[strongest], data[secondstrongest], >>> strongestantenna, secondstrongantenna), ncol =4) >>> return(value) >>> } >>> dat <- apply(mydata, 1, my.fun) >>> dat <- t(dat) >>> dat <- as.data.frame(dat, stringsAsFactors = FALSE) >>> colnames(dat) <- c("strongest", "secondstrongest", >>> "strongestantenna", "secondstrongantenna") >>> dat[ , "strongest"] <- as.numeric(dat[ , "strongest"]) >>> dat[ , "secondstrongest"] <- as.numeric(dat[ , "secondstrongest"]) >>> return(dat) >>> } >>> >>> >>> #Using your example data: >>> >>> yourdata <- structure(list(value0 = c(-13007L, -12838L, -12880L, >>> -12805L, >>> -12834L, -11068L, -12807L, -12770L, -12988L, -11779L), value60 >>> c(-11707L, >>> -13210L, -11778L, -11653L, -13527L, -11698L, -14068L, -11665L, >>> -11736L, -12873L), value120 = c(-11072L, -11176L, -11113L, -11071L, >>> -11067L, -12430L, -11092L, -11061L, -11137L, -12973L), value180 >>> c(-12471L, >>> -11799L, -12439L, -12385L, -11638L, -12430L, -11709L, -12373L, >>> -12570L, -12537L), value240 = c(-12838L, -13210L, -13089L, -11561L, >>> -13527L, -12430L, -11607L, -11426L, -13467L, -12973L), value300 >>> c(-13357L, >>> -13845L, -13880L, -13317L, -13873L, -12814L, -13025L, -12805L, >>> -13739L, -11146L)), .Names = c("value0", "value60", "value120", >>> "value180", "value240", "value300"), class = "data.frame", >>> row.names = c("1", >>> "2", "3", "4", "5", "6", "7", "8", "9", "10")) >>> >>> my.finder(yourdata) #and what you want is in a nicely labeled data >>> frame >>> >>> #A potential problem is that it is not very efficient >>> >>> #Here is a test using a matrix of 100, 000 rows >>> #sampled from the same range as your data >>> #with the same number of columns >>> >>> data.test <- matrix( >>> sample(seq(min(yourdata), max(yourdata)), size = 500000, replace >>> TRUE), >>> ncol = 5) >>> >>> system.time(my.finder(data.test)) >>> >>> #On my system I get >>> >>>> system.time(my.finder(data.test)) >>> user system elapsed >>> 2.89 0.00 2.89 >>> >>> Hope that helps, >>> >>> Josh >>> >>> >>> >>> On Fri, Jul 23, 2010 at 6:20 PM, <[256]mpward at illinois.edu> wrote: >>>> I have a data frame with a couple million lines and want to >>>> retrieve the largest and second largest values in each row, along >>>> with the label of the column these values are in. For example >>>> >>>> row 1 >>>> strongest=-11072 >>>> secondstrongest=-11707 >>>> strongestantenna=value120 >>>> secondstrongantenna=value60 >>>> >>>> Below is the code I am using and a truncated data.frame. >>>> Retrieving the largest value was easy, but I have been getting >>>> errors every way I have tried to retrieve the second largest >>>> value. I have not even tried to retrieve the labels for the >>>> value yet. >>>> >>>> Any help would be appreciated >>>> Mike >>>> >>>> >>>>> data<- >>>>> data.frame(value0, value60, value120, value180, value240, value300) >>>>> data >>>> value0 value60 value120 value180 value240 value300 >>>> 1 -13007 -11707 -11072 -12471 -12838 -13357 >>>> 2 -12838 -13210 -11176 -11799 -13210 -13845 >>>> 3 -12880 -11778 -11113 -12439 -13089 -13880 >>>> 4 -12805 -11653 -11071 -12385 -11561 -13317 >>>> 5 -12834 -13527 -11067 -11638 -13527 -13873 >>>> 6 -11068 -11698 -12430 -12430 -12430 -12814 >>>> 7 -12807 -14068 -11092 -11709 -11607 -13025 >>>> 8 -12770 -11665 -11061 -12373 -11426 -12805 >>>> 9 -12988 -11736 -11137 -12570 -13467 -13739 >>>> 10 -11779 -12873 -12973 -12537 -12973 -11146 >>>>> #largest value in the row >>>>> strongest<-apply(data, 1, max) >>>>> >>>>> >>>>> #second largest value in the row >>>>> n<-function(data)(1/(min(1/(data[1, ]-max(data[1, ]))))+ >>>>> (max(data[1, ]))) >>>>> secondstrongest<-apply(data, 1, n) >>>> Error in data[1, ] : incorrect number of dimensions >>>>> >>>> >>>> ______________________________________________ >>>> [257]R-help at r-project.org mailing list >>>> https://stat.ethz.ch/mailman/listinfo/r-help >>>> PLEASE do read the posting guide [258]http://www.R-project.org/posting-guide.html >>>> and provide commented, minimal, self-contained, reproducible code. >>>> >>> >>> >>> >>> -- >>> Joshua Wiley >>> Ph.D. Student, Health Psychology >>> University of California, Los Angeles >>> [259]http://www.joshuawiley.com/ >> >> ______________________________________________ >> [260]R-help at r-project.org mailing list >> https://stat.ethz.ch/mailman/listinfo/r-help >> PLEASE do read the posting guide [261]http://www.R-project.org/posting-guide.html >> and provide commented, minimal, self-contained, reproducible code. > > ______________________________________________ > [262]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide [263]http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. ------------------------------ Message: 34 Date: Sat, 24 Jul 2010 21:48:12 -0400 From: David Winsemius <[264]dwinsemius at comcast.net> To: David Winsemius <[265]dwinsemius at comcast.net> Cc: "[266]r-help at r-project.org list" <[267]r-help at r-project.org>, [268]mpward at illinois.edu Subject: Re: [R] Trouble retrieving the second largest value from each row of a data.frame Message-ID: <[269]9B25E777-650F-4419-92F0-9319A2B753B4 at comcast.net> Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes On Jul 24, 2010, at 9:27 PM, David Winsemius wrote: > > On Jul 24, 2010, at 8:09 PM, David Winsemius wrote: > >> >> On Jul 24, 2010, at 4:54 PM, <[270]mpward at illinois.edu> wrote: >> >>> THANKS, but I have one issue and one question. >>> >>> For some reason the "secondstrongest" value for row 3 and 6 are >>> incorrect (they are the strongest) the remaining 10 are correct?? >> >> In my run of Wiley's code I instead get identical values for rows >> 2, 5, 6. Holtman's and my solutions did not suffer from that defect, >> although mine suffered from my misreading of your request, thinking >> that you wanted the top 3. The fix is trivial >>> >>> These data are being used to track radio-tagged birds, they are >>> from automated radio telemetry receivers. I will applying the >>> following formula >>> >>> diff <- ((strongest- secondstrongest)/100) >>> bearingdiff <-30-(-0.0624*(diff**2))-(2.8346*diff) >> >> vals <- c("value0", "value60", "value120", "value180", "value240", >> "value300") >> value.str2 <- (match(yourdata$secondstrongestantenna, vals)-1)*60 Had a misspelling ... rather: match(yourdata$secondstrongantenna, vals) >> value.str1 <- (match(yourdata$strongestantenna, vals)-1)*60 >> change.ind <- abs(match(yourdata, vals) - which(match(yourdata, >> vals) ) > > OOOPs should have been > > change.ind <- abs(match(yourdata, vals) - match(yourdata, vals) ) > > >> >>> >>> A) Then the bearing diff is added to strongestantenna (value0 >>> 0degrees) if the secondstrongestatenna is greater (eg value0 and >>> value60), >> >>> B) or if the secondstrongestantenna is smaller than the >>> strongestantenna, >>> then the bearingdiff is substracted from the strongestantenna. > > yourdata$finalbearing <- with(yourdata, ifelse > (value.str2>value.str1, bearingdiff+value.str1, value.str1- > bearingdiff) ) > > >> >>> >>> C) The only exception is that if value0 (0degrees) is strongest >>> and value300(360degrees) is the secondstrongestantenna then the >>> bearing is 360-bearingdiff. >> > > yourdata$finalbearing <- with(yourdata, ifelse (strongestantenna = > "value0" & secondstrongantenna == "value300", 360- bearingdiff, > finalbearing) ); > > >>> D) Also the strongestantenna and secondstrongestantenna have to be >>> next to each other (e.g. value0 with value60, value240 with >>> value300, value0 with value300) or the results should be NA. >> >> After setting finalbearing with A, B, and C then: >> yourdata$finalbearing <- with(yourdata, ifelse ( >> change.ind <5 & change.ind > 1 , >> NA, finalbearing) ) > Better result with proper creation of value.str2: yourdata strongest secondstrongest strongestantenna secondstrongantenna finalbearing 1 -11072 -11707 value120 value60 105.48359 2 -11176 -11799 value120 value180 134.76237 3 -11113 -11778 value120 value60 106.09061 4 -11071 -11561 value120 value240 NA 5 -11067 -11638 value120 value180 135.84893 6 -11068 -11698 value0 value60 14.61868 7 -11092 -11607 value120 value240 NA 8 -11061 -11426 value120 value240 NA 9 -11137 -11736 value120 value60 104.74034 10 -11146 -11779 value300 value0 285.44272 >> >>> I have been trying to use a series of if, else statements to >>> produce these bearing, > > ifelse is the correct construct for processing vectors > > -- > David. >>> but all I am producing is errors. Any suggestion would be >>> appreciated. >> >> >>> >>> Again THANKS for you efforts. >>> >>> Mike >>> >>> ---- Original message ---- >>>> Date: Fri, 23 Jul 2010 23:01:56 -0700 >>>> From: Joshua Wiley <[271]jwiley.psych at gmail.com> >>>> Subject: Re: [R] Trouble retrieving the second largest value from >>>> each row of a data.frame >>>> To: [272]mpward at illinois.edu >>>> Cc: [273]r-help at r-project.org >>>> >>>> Hi, >>>> >>>> Here is a little function that will do what you want and return a >>>> nice output: >>>> >>>> #Function To calculate top two values and return >>>> my.finder <- function(mydata) { >>>> my.fun <- function(data) { >>>> strongest <- which.max(data) >>>> secondstrongest <- which.max(data[-strongest]) >>>> strongestantenna <- names(data)[strongest] >>>> secondstrongantenna <- names(data[-strongest])[secondstrongest] >>>> value <- matrix(c(data[strongest], data[secondstrongest], >>>> strongestantenna, secondstrongantenna), ncol =4) >>>> return(value) >>>> } >>>> dat <- apply(mydata, 1, my.fun) >>>> dat <- t(dat) >>>> dat <- as.data.frame(dat, stringsAsFactors = FALSE) >>>> colnames(dat) <- c("strongest", "secondstrongest", >>>> "strongestantenna", "secondstrongantenna") >>>> dat[ , "strongest"] <- as.numeric(dat[ , "strongest"]) >>>> dat[ , "secondstrongest"] <- as.numeric(dat[ , "secondstrongest"]) >>>> return(dat) >>>> } >>>> >>>> >>>> #Using your example data: >>>> >>>> yourdata <- structure(list(value0 = c(-13007L, -12838L, -12880L, >>>> -12805L, >>>> -12834L, -11068L, -12807L, -12770L, -12988L, -11779L), value60 >>>> c(-11707L, >>>> -13210L, -11778L, -11653L, -13527L, -11698L, -14068L, -11665L, >>>> -11736L, -12873L), value120 = c(-11072L, -11176L, -11113L, -11071L, >>>> -11067L, -12430L, -11092L, -11061L, -11137L, -12973L), value180 >>>> c(-12471L, >>>> -11799L, -12439L, -12385L, -11638L, -12430L, -11709L, -12373L, >>>> -12570L, -12537L), value240 = c(-12838L, -13210L, -13089L, -11561L, >>>> -13527L, -12430L, -11607L, -11426L, -13467L, -12973L), value300 >>>> c(-13357L, >>>> -13845L, -13880L, -13317L, -13873L, -12814L, -13025L, -12805L, >>>> -13739L, -11146L)), .Names = c("value0", "value60", "value120", >>>> "value180", "value240", "value300"), class = "data.frame", >>>> row.names = c("1", >>>> "2", "3", "4", "5", "6", "7", "8", "9", "10")) >>>> >>>> my.finder(yourdata) #and what you want is in a nicely labeled >>>> data frame >>>> >>>> #A potential problem is that it is not very efficient >>>> >>>> #Here is a test using a matrix of 100, 000 rows >>>> #sampled from the same range as your data >>>> #with the same number of columns >>>> >>>> data.test <- matrix( >>>> sample(seq(min(yourdata), max(yourdata)), size = 500000, replace >>>> TRUE), >>>> ncol = 5) >>>> >>>> system.time(my.finder(data.test)) >>>> >>>> #On my system I get >>>> >>>>> system.time(my.finder(data.test)) >>>> user system elapsed >>>> 2.89 0.00 2.89 >>>> >>>> Hope that helps, >>>> >>>> Josh >>>> >>>> >>>> >>>> On Fri, Jul 23, 2010 at 6:20 PM, <[274]mpward at illinois.edu> wrote: >>>>> I have a data frame with a couple million lines and want to >>>>> retrieve the largest and second largest values in each row, >>>>> along with the label of the column these values are in. For >>>>> example >>>>> >>>>> row 1 >>>>> strongest=-11072 >>>>> secondstrongest=-11707 >>>>> strongestantenna=value120 >>>>> secondstrongantenna=value60 >>>>> >>>>> Below is the code I am using and a truncated data.frame. >>>>> Retrieving the largest value was easy, but I have been getting >>>>> errors every way I have tried to retrieve the second largest >>>>> value. I have not even tried to retrieve the labels for the >>>>> value yet. >>>>> >>>>> Any help would be appreciated >>>>> Mike >>>>> >>>>> >>>>>> data<- >>>>>> data.frame(value0, value60, value120, value180, value240, value300) >>>>>> data >>>>> value0 value60 value120 value180 value240 value300 >>>>> 1 -13007 -11707 -11072 -12471 -12838 -13357 >>>>> 2 -12838 -13210 -11176 -11799 -13210 -13845 >>>>> 3 -12880 -11778 -11113 -12439 -13089 -13880 >>>>> 4 -12805 -11653 -11071 -12385 -11561 -13317 >>>>> 5 -12834 -13527 -11067 -11638 -13527 -13873 >>>>> 6 -11068 -11698 -12430 -12430 -12430 -12814 >>>>> 7 -12807 -14068 -11092 -11709 -11607 -13025 >>>>> 8 -12770 -11665 -11061 -12373 -11426 -12805 >>>>> 9 -12988 -11736 -11137 -12570 -13467 -13739 >>>>> 10 -11779 -12873 -12973 -12537 -12973 -11146 >>>>>> #largest value in the row >>>>>> strongest<-apply(data, 1, max) >>>>>> >>>>>> >>>>>> #second largest value in the row >>>>>> n<-function(data)(1/(min(1/(data[1, ]-max(data[1, ]))))+ >>>>>> (max(data[1, ]))) >>>>>> secondstrongest<-apply(data, 1, n) >>>>> Error in data[1, ] : incorrect number of dimensions >>>>>> >>>>> >>>>> ______________________________________________ >>>>> [275]R-help at r-project.org mailing list >>>>> https://stat.ethz.ch/mailman/listinfo/r-help >>>>> PLEASE do read the posting guide [276]http://www.R-project.org/posting-guide.html >>>>> and provide commented, minimal, self-contained, reproducible code. >>>>> >>>> >>>> >>>> >>>> -- >>>> Joshua Wiley >>>> Ph.D. Student, Health Psychology >>>> University of California, Los Angeles >>>> [277]http://www.joshuawiley.com/ >>> >>> ______________________________________________ >>> [278]R-help at r-project.org mailing list >>> https://stat.ethz.ch/mailman/listinfo/r-help >>> PLEASE do read the posting guide [279]http://www.R-project.org/posting-guide.html >>> and provide commented, minimal, self-contained, reproducible code. >> >> ______________________________________________ >> [280]R-help at r-project.org mailing list >> https://stat.ethz.ch/mailman/listinfo/r-help >> PLEASE do read the posting guide [281]http://www.R-project.org/posting-guide.html >> and provide commented, minimal, self-contained, reproducible code. > > ______________________________________________ > [282]R-help at r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide [283]http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. ------------------------------ Message: 35 Date: Sat, 24 Jul 2010 22:51:50 -0400 From: paaventhan jeyaganth <[284]paaveenthan at hotmail.com> To: r <[285]r-help at r-project.org> Subject: [R] c-statiscs 95% CI for cox regression model Message-ID: <[286]BLU140-W10DC8DB498004130C2840DB4A50 at phx.gbl> Content-Type: text/plain Dear all, how can i do the calculate the C-statistics 95% confidences interval for the cox regression model. Thanks very much for your any help. Paaveenthan _________________________________________________________________ [[elided Hotmail spam]] [[alternative HTML version deleted]] ------------------------------ Message: 36 Date: Sat, 24 Jul 2010 22:48:59 -0500 From: Dirk Eddelbuettel <[287]edd at debian.org> To: Frank E Harrell Jr <[288]f.harrell at vanderbilt.edu> Cc: "[289]r-help at r-project.org" <[290]r-help at r-project.org> Subject: Re: [R] UseR! 2010 - my impressions Message-ID: <[291]20100725034859.GA11668 at eddelbuettel.com> Content-Type: text/plain; charset=us-ascii On Sat, Jul 24, 2010 at 08:55:01AM -0500, Frank E Harrell Jr wrote: > On 07/23/2010 06:50 PM, Ravi Varadhan wrote: > I want to echo what Ravi said. The talks were terrific (thanks to > the program committee and the speakers) and Kate Mullen and her team > did an extraordinary job in putting the conference together and > running it. I am proud to have been a part of it. Thank you all! Not much to add to this, so I just leave it at "Yup!". Thanks for useR! 2010. A job well done, and then some. -- Dirk Eddelbuettel | [292]edd at debian.org | [293]http://dirk.eddelbuettel.com ------------------------------ Message: 37 Date: Sat, 24 Jul 2010 23:19:54 -0500 From: Frank E Harrell Jr <[294]f.harrell at Vanderbilt.Edu> To: paaventhan jeyaganth <[295]paaveenthan at hotmail.com> Cc: r <[296]r-help at r-project.org> Subject: Re: [R] c-statiscs 95% CI for cox regression model Message-ID: <[297]4C4BBB6A.9090400 at vanderbilt.edu> Content-Type: text/plain; charset="ISO-8859-1"; format=flowed On 07/24/2010 09:51 PM, paaventhan jeyaganth wrote: > > Dear all, > how can i do the calculate the C-statistics > 95% confidences interval for the cox regression model. > Thanks very much for your any help. > Paaveenthan install.packages('Hmisc') require(Hmisc ?rcorr.cens (there is an example at the bottom) Frank -- Frank E Harrell Jr Professor and Chairman School of Medicine Department of Biostatistics Vanderbilt University ------------------------------ Message: 38 Date: Sun, 25 Jul 2010 07:57:35 +0200 From: Michael Haenlein <[298]haenlein at escpeurope.eu> To: [299]r-help at r-project.org Subject: [R] Equivalent to go-to statement Message-ID: <[300]AANLkTimX1jOLHX6AkfzDqQEJR4LK5G_-yFfDhZK_U5_i at mail.gmail.com> Content-Type: text/plain Dear all, I'm working with a code that consists of two parts: In Part 1 I'm generating a random graph using the igraph library (which represents the relationships between different nodes) and a vector (which represents a certain characteristic for each node): library(igraph) g <- watts.strogatz.game(1, 100, 5, 0.05) z <- rlnorm(100, 0, 1) In Part 2 I'm iteratively changing the elements of z in order to reach a certain value of a certain target variable. I'm doing this using a while statement: while (target_variable < threshold) {## adapt z} The problem is that in some rare cases this iterative procedure can take very long (a couple of million of iterations), depending on the specific structure of the graph generated in Part 1. I therefore would like to change Part 2 of my code in the sense that once a certain threshold number of iterations has been achieved, the iterative process in Part 2 stops and goes back to Part 1 to generate a new graph structure. So my idea is as follows: - Run Part 1 and generate g and z - Run Part 2 and iteratively modify z to maximize the target variable - If Part 2 can be obtained in less than X steps, then go to Part 3 - If Part 2 takes more than X steps then go back to Part 1 and start again I think that R does not have a function like "go-to" or "go-back". Does anybody know of a convenient way of doing this? Thanks very much for your help, Michael [[alternative HTML version deleted]] ------------------------------ Message: 39 Date: Sat, 24 Jul 2010 23:24:40 -0700 (PDT) From: Vipul Agarwal <[301]iitkvipul at gmail.com> To: [302]r-help at r-project.org Subject: [R] Outlier Problem in Survreg Function Message-ID: <[303]1280039080326-2301422.post at n4.nabble.com> Content-Type: text/plain; charset=us-ascii Hi Everyone, I have recently started using r and working on survival analysis using the function survreg. I am facing a trange problem. One of the covariates in my analysis has outliers because of which survreg is giving incorrect results. Howevere when I am removing the outliers or scaling down the values of the covariate by a factor of 2 it is giving correct results. Below is a ditribution of the ariable and the results Min. 1st Qu. Median Mean 3rd Qu. Max. 0 30000 54500 95450 123000 1650000 Survreg Resuts survreg(formula = Surv(TIME_TO_FAILURE, CENSOR_DEFAULT) ~ ADVANCE, data = data) Coefficients: (Intercept) ADVANCE 0.000000 -6.385336 Scale= 0.9785933 Loglik(model)= -40227366 Loglik(intercept only)= -914141 Chisq= -78626451 on 1 degrees of freedom, p= 1 n=198099 (885 observations deleted due to missingness) Survreg Results after scaling down the variable by 10 survreg(formula = Surv(TIME_TO_FAILURE, CENSOR_DEFAULT) ~ ADVANCE_SCALED, data = data) Coefficients: (Intercept) ADVANCE_SCALED 4.132962e+00 -2.181577e-05 Scale= 0.9428758 Loglik(model)= -909139.4 Loglik(intercept only)= -914141 Chisq= 10003.19 on 1 degrees of freedom, p= 0 n=198099 (885 observations deleted due to missingness) Survreg Results Afte removing the outliers(5% of the obs) data <- subset(data, data$ADVANCE <= 200000) > survreg(Surv(TIME_TO_FAILURE, CENSOR_DEFAULT) ~ ADVANCE , data = data ) Call: survreg(formula = Surv(TIME_TO_FAILURE, CENSOR_DEFAULT) ~ ADVANCE, data = data) Coefficients: (Intercept) ADVANCE 4.224298e+00 -3.727171e-06 Scale= 0.9601186 Loglik(model)= -822521.9 Loglik(intercept only)= -825137.1 Chisq= 5230.49 on 1 degrees of freedom, p= 0 n=177332 (444 observations deleted due to missingness) Please let me know if someone else has faced the same problem and what is the way around to deal with it ? Should I scale down the variable or remove the outliers? -- View this message in context: [304]http://r.789695.n4.nabble.com/Outlier-Problem-in-Survreg-Function-tp2 301422p2301422.html Sent from the R help mailing list archive at Nabble.com. ------------------------------ Message: 40 Date: Sun, 25 Jul 2010 02:43:24 -0400 From: Gabor Grothendieck <[305]ggrothendieck at gmail.com> To: Michael Haenlein <[306]haenlein at escpeurope.eu> Cc: [307]r-help at r-project.org Subject: Re: [R] Equivalent to go-to statement Message-ID: <[308]AANLkTimG-oCXkzLqxpca6HnTbnErb4vbBEQKhYyOkaWs at mail.gmail.com> Content-Type: text/plain; charset=ISO-8859-1 On Sun, Jul 25, 2010 at 1:57 AM, Michael Haenlein <[309]haenlein at escpeurope.eu> wrote: > Dear all, > > I'm working with a code that consists of two parts: In Part 1 I'm generating > a random graph using the igraph library (which represents the relationships > between different nodes) and a vector (which represents a certain > characteristic for each node): > > library(igraph) > g <- watts.strogatz.game(1, 100, 5, 0.05) > z <- rlnorm(100, 0, 1) > > In Part 2 I'm iteratively changing the elements of z in order to reach a > certain value of a certain target variable. I'm doing this using a while > statement: > > while (target_variable < threshold) {## adapt z} > > The problem is that in some rare cases this iterative procedure can take > very long (a couple of million of iterations), depending on the specific > structure of the graph generated in Part 1. I therefore would like to change > Part 2 of my code in the sense that once a certain threshold number of > iterations has been achieved, the iterative process in Part 2 stops and goes > back to Part 1 to generate a new graph structure. So my idea is as follows: > > - Run Part 1 and generate g and z > - Run Part 2 and iteratively modify z to maximize the target variable > - If Part 2 can be obtained in less than X steps, then go to Part 3 > - If Part 2 takes more than X steps then go back to Part 1 and start again > > I think that R does not have a function like "go-to" or "go-back". > > Does anybody know of a convenient way of doing this? > > Thanks very much for your help, > goto's can be replaced with loops. In this case create a double loop such that the outer loop does not repeat if the inner loop finished due to reaching the target: target_variable <- -Inf while(target_variable < threshold) { ... iter <- 0 while(target_variable < threshold && iter < max_iter) { ... update iter and target_variable ... } } ------------------------------ _______________________________________________ [310]R-help at r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide [311]http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. End of R-help Digest, Vol 89, Issue 25 ************************************** [312][rKWLzcpt.zNp8gmPEwGJCA00] [@from=dllmain&rcpt=r%2Dhelp%40r%2Dproject%2Eorg&msgid=%3C20100726113540%2EH M%2E0000000000000bg%40dllmain%2Ewwl737%2Ehanmail%2Enet%3E] References 1. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 2. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-request at r-project.org 3. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-owner at r-project.org 4. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 5. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=djmuser at gmail.com 6. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=gcheer3 at gmail.com 7. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 8. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=KLCO6QX at mail.gmail.com 9. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=gcheer3 at gmail.com 10. http://r.789695.n4.nabble.com/how-to-calculate-the-product-of-every-two-elements-in-two-vectors-tp2300299p2300299.html 11. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 12. http://www.r-project.org/posting-guide.html 13. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=SetlhareL at bob.bw 14. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=at.ouchen at gmail.com 15. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 16. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=25D1D72D6E19D144AB813C9C582E16CF03F7EA27 at BOB-EXCHANGE.bob.bw 17. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-bounces at r-project.org 18. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-bounces at r-project.org 19. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 20. http://www.stata.com/help.cgi?search 21. http://www.stata.com/support/statalist/faq 22. http://www.ats.ucla.edu/stat/stata/ 23. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 24. http://www.r-project.org/posting-guide.html 25. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=SetlhareL at bob.bw 26. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=SetlhareL at bob.bw 27. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=at.ouchen at gmail.com 28. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 29. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=25D1D72D6E19D144AB813C9C582E16CF03F7EA28 at BOB-EXCHANGE.bob.bw 30. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-bounces at r-project.org 31. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-bounces at r-project.org 32. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 33. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-bounces at r-project.org 34. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-bounces at r-project.org 35. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 36. http://www.stata.com/help.cgi?search 37. http://www.stata.com/support/statalist/faq 38. http://www.ats.ucla.edu/stat/stata/ 39. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 40. http://www.r-project.org/posting-guide.html 41. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=SetlhareL at bob.bw 42. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=at.ouchen at gmail.com 43. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 44. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=25D1D72D6E19D144AB813C9C582E16CF03F7EA29 at BOB-EXCHANGE.bob.bw 45. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-bounces at r-project.org 46. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-bounces at r-project.org 47. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 48. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-bounces at r-project.org 49. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-bounces at r-project.org 50. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 51. http://www.stata.com/help.cgi?search 52. http://www.stata.com/support/statalist/faq 53. http://www.ats.ucla.edu/stat/stata/ 54. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 55. http://www.r-project.org/posting-guide.html 56. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=dwinsemius at comcast.net 57. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 58. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 59. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=D09340C5-3B64-47FA-A168-8EA347F79747 at comcast.net 60. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 61. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 62. http://www.r-project.org/posting-guide.html 63. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=assaedi76 at yahoo.com 64. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 65. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=853644.1608.qm at web45210.mail.sp1.yahoo.com 66. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=assaedi76 at yahoo.com 67. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=drbn at yahoo.com 68. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 69. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=217686.73973.qm at web113215.mail.gq1.yahoo.com 70. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=vijayamahantesh_s at dell.com 71. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 72. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=1279964891930-2300991.post at n4.nabble.com 73. http://r.789695.n4.nabble.com/Help-me-with-prediction-in-linear-model-tp2297313p2300991.html 74. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=hadley at rice.edu 75. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jdnewmil at dcn.davis.ca.us 76. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 77. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=fahim.md at gmail.com 78. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=z at mail.gmail.com 79. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jdnewmil at dcn.davis.ca.us 80. http://had.co.nz/ 81. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=f.harrell at Vanderbilt.Edu 82. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=rvaradhan at jhmi.edu 83. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 84. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 85. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=4C4AF0B5.6070300 at vanderbilt.edu 86. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=rvaradhan at jhmi.edu 87. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=cberry at tajo.ucsd.edu 88. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=marcusliu667 at yahoo.com 89. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 90. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=Pine.LNX.4.64.1007240817250.21422 at tajo.ucsd.edu 91. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=cberry at tajo.ucsd.edu 92. http://famprevmed.ucsd.edu/faculty/cberry/ 93. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=murdoch.duncan at gmail.com 94. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=cberry at tajo.ucsd.edu 95. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 96. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=4C4B0BE9.7050409 at gmail.com 97. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=cberry at tajo.ucsd.edu 98. http://famprevmed.ucsd.edu/faculty/cberry/ 99. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 100. http://www.r-project.org/posting-guide.html 101. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=bbolker at gmail.com 102. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at stat.math.ethz.ch 103. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=loom.20100724T175114-259 at post.gmane.org 104. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=brubruzao at hotmail.com 105. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 106. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=SNT111-DS23835E9F6F65737C7B0BEDBBA40 at phx.gbl 107. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=goncalves.b.b at gmail.com 108. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 109. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mattstati at yahoo.com 110. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 111. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=289386.7557.qm at web43507.mail.sp1.yahoo.com 112. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=wwwhsd at gmail.com 113. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=gcheer3 at gmail.com 114. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 115. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=XrwenTLja3ZTrwKE9USm4z at mail.gmail.com 116. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=gcheer3 at gmail.com 117. http://r.789695.n4.nabble.com/how-to-calculate-the-product-of-every-two-elements-in-two-vectors-tp2300299p2300299.html 118. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 119. http://www.r-project.org/posting-guide.html 120. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jwiley.psych at gmail.com 121. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mattstati at yahoo.com 122. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 123. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=v0shNs at mail.gmail.com 124. http://www.r-project.org/doc/bib/R-books.html 125. http://cran.r-project.org/manuals.html 126. http://tolstoy.newcastle.edu.au/R/help/04/06/0063.html 127. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mattstati at yahoo.com 128. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 129. http://www.r-project.org/posting-guide.html 130. http://www.joshuawiley.com/ 131. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=ggrothendieck at gmail.com 132. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=gcheer3 at gmail.com 133. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 134. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=dPbmg9YRPUcE2fQH_0Sqs2UJKFVJ3BS at mail.gmail.com 135. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=gcheer3 at gmail.com 136. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jcm331 at gmail.com 137. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mattstati at yahoo.com 138. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 139. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=eWg9Wcb80 at mail.gmail.com 140. http://www.amazon.com/Programming-Data-Guide-S-Language/dp/0387985034/ref=sr_1_1?ie=UTF8&s=books&qid=1279990404&sr=8-1 141. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mattstati at yahoo.com 142. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 143. http://www.r-project.org/posting-guide.html 144. http://www.r-project.org/posting-guide.html 145. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jcm0250 at unt.edu 146. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=Greg.Snow at imail.org 147. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=FarleyR at metro.net 148. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 149. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 150. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=B37C0A15B8FB3C468B5BC7EBC7DA14CC633A53DF11 at LP-EXMBVS10.CO.IHC.COM 151. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=greg.snow at imail.org 152. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-bounces at r-project.org 153. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 154. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 155. http://www.r-project.org/posting- 156. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=shabnambioinfo at gmail.com 157. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 158. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=Hzat_FD6jpZF6 at mail.gmail.com 159. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=Greg.Snow at imail.org 160. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=babyfoxlove1 at sina.com 161. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=babyfoxlove1 at sina.com 162. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 163. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 164. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=B37C0A15B8FB3C468B5BC7EBC7DA14CC633A53DF25 at LP-EXMBVS10.CO.IHC.COM 165. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=greg.snow at imail.org 166. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help-bounces at r-project.org 167. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=babyfoxlove1 at sina.com 168. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 169. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 170. http://www.r-project.org/posting- 171. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=AbdiA at si.edu 172. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 173. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 174. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=97679C0A11332E48A01E0D463E8B3FF103AF0CE275 at SI-MSEV02.US.SINET.SI.EDU 175. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=abdia at si.edu 176. http://nationalzoo.si.edu/SCBI/ConservationGIS/ 177. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mazatlanmexico at yahoo.com 178. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at stat.math.ethz.ch 179. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=418059.32636.qm at web56602.mail.re3.yahoo.com 180. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=wwwhsd at gmail.com 181. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mazatlanmexico at yahoo.com 182. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at stat.math.ethz.ch 183. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=AANLkTimd6DLagDiHKbFuF3hE5dES4d6DmC5T50Vn25N3 at mail.gmail.com 184. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mazatlanmexico at yahoo.com 185. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 186. http://www.r-project.org/posting-guide.html 187. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 188. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jwiley.psych at gmail.com 189. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 190. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=20100724155451.CHG28413 at expms6.cites.uiuc.edu 191. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jwiley.psych at gmail.com 192. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 193. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 194. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 195. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 196. http://www.r-project.org/posting-guide.html 197. http://www.joshuawiley.com/ 198. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=phhs80 at gmail.com 199. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 200. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=oH at mail.gmail.com 201. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=hw_joyce_cn at hotmail.com 202. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jholtman at gmail.com 203. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mazatlanmexico at yahoo.com 204. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at stat.math.ethz.ch 205. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=3P at mail.gmail.com 206. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mazatlanmexico at yahoo.com 207. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 208. http://www.r-project.org/posting-guide.html 209. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=zachmohr at gmail.com 210. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 211. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=AANLkTinzngqO3S_p-E25N3TTWEdiQBBS4CvzErgywj1p at mail.gmail.com 212. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=2300793-1751019155-246278 at n4.nabble.com 213. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=2B2300793-1751019155-246278 at n4.nabble.com 214. http://r.789695.n4.nabble.com/glm-prediction-of-a-factor-with-several-levels-tp2300793p2300793.html 215. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=zachmohr at gmail.com 216. http://r.789695.n4.nabble.com/glm-prediction-of-a-factor-with-several-levels-tp2300793p2301324.html 217. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=dwinsemius at comcast.net 218. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 219. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 220. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=52EA484F-C066-4ACC-B5BC-1A3A20876D9E at comcast.net 221. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 222. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jwiley.psych at gmail.com 223. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 224. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 225. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 226. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 227. http://www.r-project.org/posting-guide.html 228. http://www.joshuawiley.com/ 229. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 230. http://www.r-project.org/posting-guide.html 231. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jwiley.psych at gmail.com 232. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 233. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 234. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=fa9gCQLuYBAaMS_P8w at mail.gmail.com 235. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=dwinsemius at comcast.net 236. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 237. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jwiley.psych at gmail.com 238. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 239. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 240. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 241. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 242. http://www.r-project.org/posting-guide.html 243. http://www.joshuawiley.com/ 244. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 245. http://www.r-project.org/posting-guide.html 246. http://www.joshuawiley.com/ 247. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=dwinsemius at comcast.net 248. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 249. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 250. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 251. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=2B19FDC3-4358-4731-87C7-89399E5DD75E at comcast.net 252. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 253. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jwiley.psych at gmail.com 254. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 255. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 256. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 257. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 258. http://www.r-project.org/posting-guide.html 259. http://www.joshuawiley.com/ 260. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 261. http://www.r-project.org/posting-guide.html 262. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 263. http://www.r-project.org/posting-guide.html 264. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=dwinsemius at comcast.net 265. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=dwinsemius at comcast.net 266. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 267. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 268. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 269. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=9B25E777-650F-4419-92F0-9319A2B753B4 at comcast.net 270. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 271. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=jwiley.psych at gmail.com 272. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 273. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 274. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=mpward at illinois.edu 275. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 276. http://www.r-project.org/posting-guide.html 277. http://www.joshuawiley.com/ 278. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 279. http://www.r-project.org/posting-guide.html 280. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 281. http://www.r-project.org/posting-guide.html 282. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 283. http://www.r-project.org/posting-guide.html 284. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=paaveenthan at hotmail.com 285. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 286. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=BLU140-W10DC8DB498004130C2840DB4A50 at phx.gbl 287. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=edd at debian.org 288. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=f.harrell at vanderbilt.edu 289. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 290. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 291. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=20100725034859.GA11668 at eddelbuettel.com 292. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=edd at debian.org 293. http://dirk.eddelbuettel.com/ 294. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=f.harrell at Vanderbilt.Edu 295. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=paaveenthan at hotmail.com 296. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 297. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=4C4BBB6A.9090400 at vanderbilt.edu 298. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=haenlein at escpeurope.eu 299. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 300. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=AANLkTimX1jOLHX6AkfzDqQEJR4LK5G_-yFfDhZK_U5_i at mail.gmail.com 301. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=iitkvipul at gmail.com 302. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 303. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=1280039080326-2301422.post at n4.nabble.com 304. http://r.789695.n4.nabble.com/Outlier-Problem-in-Survreg-Function-tp2301422p2301422.html 305. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=ggrothendieck at gmail.com 306. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=haenlein at escpeurope.eu 307. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=r-help at r-project.org 308. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=AANLkTimG-oCXkzLqxpca6HnTbnErb4vbBEQKhYyOkaWs at mail.gmail.com 309. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=haenlein at escpeurope.eu 310. http://mail2.daum.net/hanmail/mail/MailComposeFrame.daum?TO=R-help at r-project.org 311. http://www.r-project.org/posting-guide.html 312. mailto:dllmain at hanmail.net
나여나
2010-Jul-26 08:32 UTC
[R] After writing data in MMF using SEXP structure, can i reference in R?
Hi all, After writing data in MMF(Memory Map File) using SEXP structure, can i reference in R? If input data is larger than 2GB, Can i reference MMF Data in R? my work environment : R version : 2.11.1 OS : WinXP Pro sp3 Thanks and best regards. Park, Young-Ju from Korea. [1][rKWLzcpt.zNp8gmPEwGJCA00] [@from=dllmain&rcpt=r%2Dhelp%40r%2Dproject%2Eorg&msgid=%3C20100726173212%2EH M%2E0000000000000bk%40dllmain%2Ewwl737%2Ehanmail%2Enet%3E] References 1. mailto:dllmain at hanmail.net