similar to: R --nsize 2M runs havoc (under linux)

Displaying 20 results from an estimated 4000 matches similar to: "R --nsize 2M runs havoc (under linux)"

2001 Feb 22
1
R-1.2.0: bug report on fn table
> table("NA") numeric(0) > table("NaN") numeric(0) This should probably return something different, like in > table("nan") nan 1 > or do I misunderstand something? regards Joerg. -- Dr. Joerg Kindermann Knowledge Discovery Team GMD - German National Research Center for Information Technology - phone:
1999 Aug 26
0
use dyn.load to redefine R internal fns?
I would like to redefine some R print functions (Rprintf for example) so that they don't print to stdout, but save the output in a global string array. I can compile and link my definition of Rprintf into a library and load it with dyn.load. I don't get an error message, but afterwards the original function is still valid, my new version isn't executed. Is there a way to change R
2000 Feb 23
1
Version 0.90.1 bug report on matrix indexing
Hi , R Version 0.90.1 on Solaris2.5 and Suse Linux 6.[1,3] crashes some time after a matrix row or column has been addressed via an incorrect row/colname: R : Copyright 1999, The R Development Core Team Version 0.90.1 (December 15, 1999) R is free software and comes with ABSOLUTELY NO WARRANTY. You are welcome to redistribute it under certain conditions. Type "?license" or
2000 Jan 17
1
dyn.load problem
Hi, I'm trying to write modified versions of R save/load functions which read from/write to a char array instead of a file pointer. I want to compile my functions into a library and dyn.load them. Of course the new functions mySave/Load will use some of the internal R functions used by save/load, for example "MarkSave" (to be found in .../src/main/saveload.c). I declared them as
2001 Jan 08
0
R-1.2.0: gdb doesn't find functions in shared lib
I'm trying to debug some C functions I've written for use with R. Until recently debugging them with gdb was no problem. But since I switched to R-1.2.0 (from R-1.0.1) gdb can no longer find the entry points of my functions. It either pretends to set break points but doesn't stop there or it only stops at the end of the function. R version R-1.2.0 gdb version GNU gdb 4.18 gcc version
2000 Nov 09
3
maximum of nsize=20000k ??
Dear R-ers, somehow it is not possible to increase nsize to more than 20000k. When I specify e.g. > R --vsize=10M --nsize=21000K the result is: free total (Mb) Ncells 99658 350000 6.7 Vcells 1219173 1310720 10.0 Maybe I have overlooked s.th.... Marcus -- +------------------------------------------------------- | Marcus Eger | E-Mail: eger.m at gmx.de (NEW) |
1999 May 15
2
vsize and nsize
I am running R version ??? under Redhat 5.2. It seems as though the --nsize object has no effct on the size of the allocated Ncells as determined using gc(). Yes, I have that much data.... That is if I envoke R with R --vsize 100 --nsize 5000000 then type gc() I get free total Ncells 92202 200000 Vcells 12928414 13107200 Thanks Tony Long Ecology and Evolutionary Biology Steinhaus
1999 Apr 12
3
--nsize and --vsize
Martin M has suggested I widen this discussion to R-devel, and > I agree that we should increase them, > but I'm not sure at all about the amount. > > The default could even depend on the architecture (via "./configure").. Views, please. ------------- Begin Forwarded Message ------------- Is is not time we increased the defaults a bit? As the base gets bigger I hit
2015 Jan 15
2
default min-v/nsize parameters
Just wanted to start a discussion on whether R could ship with more appropriate GC parameters. Right now, loading the recommended package Matrix leads to: > library(Matrix) > gc() used (Mb) gc trigger (Mb) max used (Mb) Ncells 1076796 57.6 1368491 73.1 1198505 64.1 Vcells 1671329 12.8 2685683 20.5 1932418 14.8 Results may vary, but here R needed 64MB of N cells and 15MB
2004 Jul 20
1
--max-vsize and --max-nsize linux?
Hi, somtimes i have trivial recodings like this: > dim(tt) [1] 252382 98 system.time(for(i in 2:length(tt)){ tt[,i][is.na(tt[,i])] <- 0 }) ...and a win2000(XP2000+,1GB) machine makes it in several minutes, but my linux notebook (XP2.6GHZ,512MB) don't get success after some hours. I recognize that the cpu load is most time relative small, but the hardisk
2015 Jan 18
2
default min-v/nsize parameters
On Thu, Jan 15, 2015 at 3:55 PM, Michael Lawrence <lawrence.michael at gene.com> wrote: > Just wanted to start a discussion on whether R could ship with more > appropriate GC parameters. I've been doing a number of similar measurements, and have come to the same conclusion. R is currently very conservative about memory usage, and this leads to unnecessarily poor performance on
2015 Jan 20
1
default min-v/nsize parameters
>>>>> Peter Haverty <haverty.peter at gene.com> >>>>> on Mon, 19 Jan 2015 08:50:08 -0800 writes: > Hi All, This is a very important issue. It would be very > sad to leave most users unaware of a free speedup of this > size. These options don't appear in the R --help > output. They really should be added there. Indeed,
1999 Oct 13
3
main/character.c (et.al): dangerous AllocBuffer()
I was hit by ugly crashes of R, when I tried to read big data sets ("volcano"). So I looked into the code and found the following in character.c (triggered by substr()): I assume that the helper function AllocBuffer() shall facilitate an economic memory management. But the use of realloc() in the else-branch does not conform to ANSI and may hit you with certain compilers. When called
1999 Oct 13
3
main/character.c (et.al): dangerous AllocBuffer()
I was hit by ugly crashes of R, when I tried to read big data sets ("volcano"). So I looked into the code and found the following in character.c (triggered by substr()): I assume that the helper function AllocBuffer() shall facilitate an economic memory management. But the use of realloc() in the else-branch does not conform to ANSI and may hit you with certain compilers. When called
2001 Aug 22
1
Huge workspace cannot be opened
Hi everyone, I have a problem that some people may have already encountered but i did not find the solution yet. As I use R to simulate several arrays of data, my workspace is now 35Mb big and I cannot launch R with it. An "xdr real data read error occured" and R tells me to delete .RData or increase memory. I WON'T delete this file and changing the max-nsize to 40600k did not
2000 Sep 07
6
Writing a tabel in LaTeX-format
I have a vague recollection of seeing a R-function that will output a table or matix in a format suitable for the tabular enviroment in LaTeX. But I cannot find it. Any hints? Bendix ---------------------- Bendix Carstensen Senior Statistician Steno Diabetes Centre Niels Steensens Vej 2 DK-2820 Gentofte Denmark tel: +45 44 43 87 38 mob: +45 28 25 87 38 fax: +45 44 43 73 13 bxc at novo.dk
2023 May 25
1
data.frame with a column containing an array
So is this an expected behavior or is it a bug which should be reported somewhere else? Thanks! Georg? ? ? Gesendet:?Dienstag, 09. Mai 2023 um 19:28 Uhr Von:?"Bert Gunter" <bgunter.4567 at gmail.com> An:?"Georg Kindermann" <Georg.Kindermann at gmx.at> Cc:?"Rui Barradas" <ruipbarradas at sapo.pt>, r-help at r-project.org Betreff:?Re: [R] data.frame
2023 May 25
3
data.frame with a column containing an array
I really don't know. I would call it a request for extended capabilities of [.data.frame, rather than a feature or bug. But maybe wiser heads than mine who monitor this list can sort it out. -- Bert On Wed, May 24, 2023 at 8:52?PM Georg Kindermann <Georg.Kindermann at gmx.at> wrote: > So is this an expected behavior or is it a bug which should be reported > somewhere else? >
2015 Jan 17
0
default min-v/nsize parameters
Martin Morgan discussed this a year or so ago and as I recall bumped up these values to the current defaults. I don't recall details about why we didn't go higher -- maybe Martin does. I suspect the main concern would be with small memory machines in student labs and less developed countries. If there was a way on all platforms to identify how much memory is available that might help to
2004 Oct 07
5
Question restricting ssh access for some users only
I've used ssh as a secure telnet up to now but done little else with it. The FreeBSD machines I look after on our internet-facing network all have one account which I connect to for administration. I've set up /etc/hosts.allow on all the machines to only allow ssh from a limited internal network range. Now I want to create a new account on one machine which will be accessible from the