similar to: heap memory

Displaying 20 results from an estimated 30000 matches similar to: "heap memory"

1998 Nov 26
1
heap memory exhausted
Hi I always have following error message when I try to read a big ascii-file: > inzp<-read.data() > Error: heap memory (1953 Kb) exhausted [needed 0 Kb more] read.data() is a small function that reads the ascii-file. When I cut the ascii-file to a small one I don't have this problem. Can I extend this 'heap memory' for reading big data files? How? I'm working with R on a
1999 Jan 21
2
scoping problem?
Dear R-helpers: (this is part of a bigger program) the following fails as a function, but runs OK if we comment out the fnfn_ function() line. Any hint would be appreciated. -Yudi- R : Copyright 1998, The R Development Core Team Version 0.63.0 Beta (Nov 13, 1998) -- on WIndows3.11 fnfn _ function (m=10,n=10,spar=2) { fn _ function(u,v){ uc_ u-floor(m/2)-1 vc_ v-floor(n/2)-1
2000 Mar 03
1
tapply, sorting and the heap
howdy gurus, I'm new and green and I was hoping for a tiny bit of your expertise. I'm running out of virtual memory (heap?) when summing using tapply. I've already used --vsize=90M on my hpux machine. (details below) Can I pre-sort or something to prevent my error? thanks, John Strumila john.strumila at corpmail.telstra.com.au > gc()["Vcells","total"] [1]
2000 Feb 11
1
astonishing memory phenomenon
I have a question concerning memory. I understood that R takes a fixed amount of memory at startup (which I can influence with --vsize --nsize) and that gc() shows the memory still free of the total memory reserved for R. However, if I create a long vector of character data, gc() only seem to reflect the space needed for a vector of pointers to char, the space used for the character data itself
2010 May 14
2
[Bug 1771] New: sftp heap memory leak
https://bugzilla.mindrot.org/show_bug.cgi?id=1771 Summary: sftp heap memory leak Product: Portable OpenSSH Version: 5.5p1 Platform: All OS/Version: All Status: NEW Severity: normal Priority: P2 Component: sftp AssignedTo: unassigned-bugs at mindrot.org ReportedBy: anicka at suse.cz
1999 Apr 27
2
Memory management
Dear all, I don't get it: First of all, the help doesn't say what are the memory limits of R. Say, what's the max heap size for instance ???? Secondly, I invoke R with the following commands each time: rgui --vsize 30M --nsize 1000K rgui --vsize 30M --nsize 2000K rgui --vsize 30M --nsize 3000K rgui --vsize 30M --nsize 4000K I try to open a matrix 8000x8000 by issuing
1998 Nov 26
1
Saving memory usage -- .C(....., DUP = FALSE) danger?
Just found out [R 0.63, standard -v -n] : > rm(list=ls()) > gc() free total Ncells 96538 200000 Vcells 214008 250000 > hist(runif(50000)) Error: heap memory (1953 Kb) exhausted [needed 390 Kb more] which is a bit astonishing given that I still have room for 214000 double's > u1 <- runif(50000) > u2 <- runif(50000) > gc()
1998 Nov 19
0
R data import problems
Excuse me, if the following is FAQ or nonsense, I'm new to R (and the list)! After trying to import a textfile of 184897 Bytes, R (v0.63.0 on Linux 2.0.35/DLD 5.4) gives the following error message: > read.table ("db.out") Error: heap memory (1953 Kb) exhausted [needed 150 Kb more] > Is there a possibility to increase
1999 Jun 04
1
dealing with large objects -- memory wasting ?
Consider the following: > gcinfo(TRUE) [1] FALSE > rm(list=ls()); gc() free total Ncells 135837 250000 Vcells 747306 786432 ------ > n <- 10000; p <- 20; X <- matrix(rnorm(n*p), n,p); gc() Garbage collection [nr. 23]... 135839 cons cells free (54%) 4275 Kbytes of heap free (69%) free total Ncells 135829 250000 Vcells
2001 Apr 02
1
Run out of memory
I am trying to use R to cluster 7129 samples, my data set is a 7129 x 38 matrix, when I trying to get the distance matrix using function dist( ), the memory exhausted, and I tried to set the memory when I run R by R --vsize=250M --nsize=1000k no matter what I set for vsize the result is the same, it says: Error: heap memory (256000kb) exhausted [need 198498Kb more]
2014 Mar 19
2
[LLVMdev] getElapsedWallTime unnecessary heap allocation and memory leak
> From: llvmdev-bounces at cs.uiuc.edu [mailto:llvmdev-bounces at cs.uiuc.edu] > On Behalf Of Bryan Keiren > Subject: [LLVMdev] getElapsedWallTime unnecessary heap allocation and memory leak > In the file \lib\Support\Process.cpp on line 60, it seems as though an > unnecessary heap allocation and memory leak occurs. > static TimeValue getElapsedWallTime() { >  static
1999 Apr 26
1
random sequence
Dear All, How can I get the same stream of random numbers in R and S? I know one of the generators in R is Super-Duper also used in S-plus, but the .Random.seed's are set differently. I tried to re-exrpess that in S-plus into two seeds according to the formula in VR, but I can't get the same random stream in R. Thanks, ------------------------------ Yudi Pawitan: yudi at ucd.ie
1999 Apr 26
1
random sequence
Dear All, How can I get the same stream of random numbers in R and S? I know one of the generators in R is Super-Duper also used in S-plus, but the .Random.seed's are set differently. I tried to re-exrpess that in S-plus into two seeds according to the formula in VR, but I can't get the same random stream in R. Thanks, ------------------------------ Yudi Pawitan: yudi at ucd.ie
2008 Feb 06
1
Re-map Portion of Ruby Heap Holding Optree to Save Child Server Memory?
I wonder if it is possible to reduce the memory footprint of a pack of 6 mongrels by about 70 meg. In reading Hongli''s blog about his revisions to Ruby GC (last Monday Matz if he could use Hongli''s work), I was wondering at the size (in RSS) of all the mongrels in a cluster. It has always seemed just too big for my C-binary intuition. Why is so much of the memory of the parent
2004 Feb 04
0
help(Memory) [forwarded message]
I can't understand that people still send things like this to R-core... ------- start of forwarded message ------- From: Tineke Casneuf <ticas at psb.ugent.be> Sender: r-core-bounces at stat.math.ethz.ch To: R-core at r-project.org Subject: help(Memory) Date: Wed, 04 Feb 2004 13:39:32 +0100 Dear, I am trying to find a appropriate package to analyse gene expression data from DNA
2014 Mar 19
2
[LLVMdev] getElapsedWallTime unnecessary heap allocation and memory leak
On Wed, Mar 19, 2014 at 6:18 PM, Bryan Keiren < bryan.keiren at guerrilla-games.com> wrote: > We are indeed trying to completely clean the heap before exiting main(). > Which means that you either don't have threads, or you join all threads before main exits. Is that the case? Good for you if so! > > > -----Original Message----- > From: Caldarale, Charles R
2000 Feb 10
0
Re: your mail about Memory on Windows95
>I use WinNT. You have to "launch" R from a DOS-shell window; you 1st >change to the directory where you have RGui.exe, then type : > Rgui --vsize 15M --nsize 1000k > >and R opens with increased memore size. It works here with my NT box, and >will probably do with Win95 (though I think that the DOS-shells are not >strictly similar in both OSs but this may not be a
2004 Jun 14
0
[LLVMdev] Memory Alignment, Heap allocation.
Tobias Nurmiranta wrote: > Hi, Chris and others can give you better ideas on the ideal ways to implement what you want, but I'll give some ideas/answers for now. > > 1. > > A small question: How do I ensure memory alignment? I want all malloced > memory, globals and functions to be 4-byte aligned. Does llvm have any > ".align" keyword? No, LLVM does not
2011 Jul 19
2
[LLVMdev] Heap memory allocation
I'm implementing a lazy functional language and I need to allocate nodes on the heap. From what I've read LLVM supports only the 'alloca' instruction which allows to allocate memory on the stack. Is this true and and if so, what is the preferred way to solve my problem? -- Piotr Kaleta -------------- next part -------------- An HTML attachment was scrubbed... URL:
1999 Nov 25
1
segfault in garbage collection (PR#344)
The following statements yield a seg.fault: R --vanilla --nsize 500K x <- rep(letters,10000) f <- function(x) {z<-paste("\"",x,"\"",sep=""); z} y <- f(x) Segmentation fault If a turn gcinfo on, I get Garbage collection [nr. 1]... 387529 cons cells free (75%) 3807 Kbytes of heap free (62%) Garbage collection [nr. 2]... 273868 cons cells free