similar to: Memory Problems with a Simple Bootstrap - Part II

Displaying 20 results from an estimated 700 matches similar to: "Memory Problems with a Simple Bootstrap - Part II"

2007 Apr 27
1
bootstrap
Dear All, I would like to use a nonparametric bootstrap to calculate the confidence intervals for the 5% and 95% quantiles using boot.ci. As you know, boot.ci requires the use of boot to generate bootstrap replicates for my statistic. However this last function doesn't work in my case because I am missing something. Here is an example y <- rnorm(100) Quantile <-
2008 May 01
1
Need Help! (Bug Report)
Need Help!! Server : AIX 5.3 Compile Error Message configure --------------------------------- checking jfs/quota.h usability... no checking jfs/quota.h presence... yes configure: WARNING: jfs/quota.h: present but cannot be compiled configure: WARNING: jfs/quota.h: check for missing prerequisite headers? configure: WARNING: jfs/quota.h: see the Autoconf documentation
2018 Jun 01
4
Regroup and create new dataframe
Hello folks, I have a big project to work on and the dataset is classified so I am just going to use my own example so everyone can understand what I am targeting. Let's take Target as an example: We consider three brands of tape: Target brand, 3M and Avery. The original data frame has 4 columns: Year of Record, Product_Name(which contains three brands of tape), Sales, and Region. I want to
2008 Feb 18
2
Custom Plot - means, SD & 5th-95th% (Plotmeans or Boxplot)?
Any help with this problem would be greatly appreciated: I need to produce a custom plot i haven't come across in R. Basically, I want to show means, 1st standard deviation and 5th and 95th percentiles visually, using something resembling a boxplot. Is it possible to completely customize a boxplot so that it shows means as the bar (instead of, not as well as medians), standard deviations at
2011 Apr 17
3
Box plot with 5th and 95th percentiles instead of 1.5 * IQR: problems implementing an existing solution...
Hi all, I'm just getting started with R and I would appreciate some help. I'm having trouble creating a boxplot with whiskers at the 95th and 5th percentiles instead of at 1.5 * IQR. I have read the relevant documentation, and checked existing mails on this topic. I found a small modification that should work : https://stat.ethz.ch/pipermail/r-help/2001-November/016817.html and tried to
2006 Jul 24
2
How to obtain 95th percentile of a normal distribution of a continuous variable
Hi, How do I get R to output the 95% cutoff from a distribution of a continous variable? summary() only displays a few statistics.... Thanks!
2013 Aug 29
23
[PATCH] Btrfs: optimize key searches in btrfs_search_slot
When the binary search returns 0 (exact match), the target key will necessarily be at slot 0 of all nodes below the current one, so in this case the binary search is not needed because it will always return 0, and we waste time doing it, holding node locks for longer than necessary, etc. Below follow histograms with the times spent on the current approach of doing a binary search when the
2010 Apr 05
3
Creating R packages, passing by reference and oo R.
Dear All, I would like some advice on creating R packages, passing by reference and oo R. I have created a package that is neither elegant nor extensible and rather cumbersome (it works). I would like to re write the code to make the package distributable (should it be of interest) and easy to maintain. The package is for Bayesian model determination via a reversible jump algorithm and has
2004 Aug 31
4
More efficient matrix computation
I have a 20x3 matrix as follows: > m <- replicate(3, matrix(rnorm(20),20,1)) I need to compute, say, 95th and 99th percentiles of each column such that the resulting matrix becomes 2x3 with each row representing the respective percentile. My "best effort" is to compute one column at a time as follows: > quantile(m[,1], c(0.95, 0.99)) To do the same for columns 2 and 3, I
2012 Nov 08
3
difference percentile R vs SPSS
Dear list, I am calculating the 95th percentile of a set of values with R and with SPSS In R: > normal200<-rnorm(200,0,1) > qnorm(0.95,mean=mean(normal200),sd=sd(normal200),lower.tail =TRUE) [1] 1.84191 In SPSS, if I use the same 200 values and select Analyze -> Descriptive Statistics -> Frequencies and under "Statistics", I type in '95' under Percentiles,
2008 Jan 30
1
Understanding an R improvement that already occurred.
I was surprised to observe the following difference between 2.4.1 and 2.6.0 after a long overdue upgrade a few months ago of our departmental server. It wasn't a bug fix, but a subtle improvement. Here's the simplest example I could create. The size is excessive, on the order of the Netflix Competition data. The integer matrix is about 1.12 GB, and if coerced to numeric it is 2.24 GB.
2010 Apr 05
1
new to R, analysis of latency data
Hi, I'd like to move from excel to R because our dataset are so large. Here's what my data looks like: Transaction Rate Run# Transaction Type Location Latency in Seconds 10 1 Order A 0 10 1 Order B
2005 Nov 15
1
cannot.allocate.memory.again and 32bit<--->64bit
hello! ------ i use 32bit.Linux(SuSe)Server, so i'm limited with 3.5Gb of memory i demonstrate, that there is times to times a problem with allocating of objects of large size, for example 0.state (no objects yet created) ------------------------------------ > gc() used (Mb) gc trigger (Mb) max used (Mb) Ncells 162070 4.4 350000 9.4 350000
2010 Dec 17
3
box-and-whisker plots based on summary not data
Hi, Is it possible to produce box-and-whisker plots given that I have the median, interquartile and 5/95th centile values, but not the data from which they come? It seems that it ought to be possible to coerce bxp to do what I want, but I can't quite see how. Thanks, Matthew -- Matthew Vernon, Research Fellow Ecology and Epidemiology Group, University of Warwick
2006 May 16
2
Large database help
Hello all. I have a large .txt file whose variables are fixed-columns, ie, variable V1 goes from columns 1 to 7, V2 from 8 to 23 etc. This is a 60GB file with 90 variables and 60 million observations. I'm working with a Pentium 4, 1GB RAM, Windows XP Pro. I tried the following code just to see if I could work with 2 variables but it seems not possible: R : Copyright 2005, The R Foundation
2007 Mar 28
2
Suggestion for memory optimization and as.double() with friends
Hi, when doing as.double() on an object that is already a double, the object seems to be copied internally, doubling the memory requirement. See example below. Same for as.character() etc. Is this intended? Example: % R --vanilla > x <- double(1e7) > gc() used (Mb) gc trigger (Mb) max used (Mb) Ncells 234019 6.3 467875 12.5 350000 9.4 Vcells 10103774 77.1
2005 Dec 14
2
The fastest way to select and execute a few selected functions inside a function
Dear useRs? I have the following problem! I have a function that calls one or more functions, depending on the input parameters. I am searching for the fastest way to select and execute the selected functions and return their results in a list. The number of possible functions is 10, however usually only 2 are selected (although sometimes more, even all). For examples, if I have function
2008 Mar 24
1
Cannot allocate large vectors (running out of memory?)
Hi. As shown in the simplified example below, I'm having trouble allocating memory for large vectors, even though it would appear that there is more than enough memory available. That is, even with a memory limit of 1500 MB, R 2.6.1 (Win) will allocate memory for a first vector of 285 MB, but not for a second vector of the same size. Forcing garbage collection does not seem
2001 Nov 25
2
Boxplots using percentiles?
The standard R boxplot appears to use quartiles to determine the height of the rectangles and a range parameter - RNG - (default=1.5 I think) that determines the length of the whiskers as <= RNG x Interquartile Range. Is it possible to instead define the range as extending to the 95th percentile? If so, how would this be done? nb, I'm plotting multiple boxplots on a single chart so a
2010 Jun 24
4
OT: Bandwidth calculations
Hi, I know some of you are very experienced as to the working of networks. I wondered whether there is some accepted way of determining bandwidth needs based on the network traffic over time. For example, looking at the figures for the network traffic through the server interface, we have hourly, daily and monthly figures. If everything were linear, taking the hourly figure and dividing it by