Displaying 20 results from an estimated 4000 matches similar to: "R --nsize 2M runs havoc (under linux)"
2001 Feb 22
1
R-1.2.0: bug report on fn table
> table("NA")
numeric(0)
> table("NaN")
numeric(0)
This should probably return something different, like in
> table("nan")
nan
1
>
or do I misunderstand something?
regards
Joerg.
-- Dr. Joerg Kindermann Knowledge Discovery Team
GMD - German National Research Center for Information Technology -
phone:
1999 Aug 26
0
use dyn.load to redefine R internal fns?
I would like to redefine some R print functions (Rprintf for example) so
that they don't print to stdout, but save the output in a global string
array. I can compile and link my definition of Rprintf into a library and
load it with dyn.load. I don't get an error message, but afterwards the
original function is still valid, my new version isn't executed.
Is there a way to change R
2000 Feb 23
1
Version 0.90.1 bug report on matrix indexing
Hi ,
R Version 0.90.1 on Solaris2.5 and Suse Linux 6.[1,3] crashes some time
after a matrix row or column has been addressed via an incorrect
row/colname:
R : Copyright 1999, The R Development Core Team
Version 0.90.1 (December 15, 1999)
R is free software and comes with ABSOLUTELY NO WARRANTY.
You are welcome to redistribute it under certain conditions.
Type "?license" or
2000 Jan 17
1
dyn.load problem
Hi,
I'm trying to write modified versions of R save/load functions which read
from/write to a char array instead of a file pointer. I want to compile
my functions into a library and dyn.load them. Of course the new functions
mySave/Load will use some of the internal R functions used by save/load,
for example "MarkSave" (to be found in .../src/main/saveload.c). I
declared them as
2001 Jan 08
0
R-1.2.0: gdb doesn't find functions in shared lib
I'm trying to debug some C functions I've written for use with R. Until
recently debugging them with gdb was no problem. But since I switched to
R-1.2.0 (from R-1.0.1) gdb can no longer find the entry points of my
functions. It either pretends to set break points but doesn't stop there or
it only stops at the end of the function.
R version R-1.2.0
gdb version GNU gdb 4.18
gcc version
2000 Nov 09
3
maximum of nsize=20000k ??
Dear R-ers,
somehow it is not possible to increase nsize to more than
20000k. When I specify e.g.
> R --vsize=10M --nsize=21000K
the result is:
free total (Mb)
Ncells 99658 350000 6.7
Vcells 1219173 1310720 10.0
Maybe I have overlooked s.th....
Marcus
--
+-------------------------------------------------------
| Marcus Eger
| E-Mail: eger.m at gmx.de (NEW)
|
1999 May 15
2
vsize and nsize
I am running R version ??? under Redhat 5.2. It seems as though the
--nsize object has no effct on the size of the allocated Ncells as
determined using gc(). Yes, I have that much data....
That is if I envoke R with
R --vsize 100 --nsize 5000000
then type
gc()
I get
free total
Ncells 92202 200000
Vcells 12928414 13107200
Thanks
Tony Long
Ecology and Evolutionary Biology
Steinhaus
1999 Apr 12
3
--nsize and --vsize
Martin M has suggested I widen this discussion to R-devel, and
> I agree that we should increase them,
> but I'm not sure at all about the amount.
>
> The default could even depend on the architecture (via "./configure")..
Views, please.
------------- Begin Forwarded Message -------------
Is is not time we increased the defaults a bit? As the base gets bigger
I hit
2015 Jan 15
2
default min-v/nsize parameters
Just wanted to start a discussion on whether R could ship with more
appropriate GC parameters. Right now, loading the recommended package
Matrix leads to:
> library(Matrix)
> gc()
used (Mb) gc trigger (Mb) max used (Mb)
Ncells 1076796 57.6 1368491 73.1 1198505 64.1
Vcells 1671329 12.8 2685683 20.5 1932418 14.8
Results may vary, but here R needed 64MB of N cells and 15MB
2004 Jul 20
1
--max-vsize and --max-nsize linux?
Hi,
somtimes i have trivial recodings like this:
> dim(tt)
[1] 252382 98
system.time(for(i in 2:length(tt)){
tt[,i][is.na(tt[,i])] <- 0
})
...and a win2000(XP2000+,1GB) machine makes it in several minutes, but
my linux notebook (XP2.6GHZ,512MB) don't get success after some hours.
I recognize that the cpu load is most time relative small, but the hardisk
2015 Jan 18
2
default min-v/nsize parameters
On Thu, Jan 15, 2015 at 3:55 PM, Michael Lawrence
<lawrence.michael at gene.com> wrote:
> Just wanted to start a discussion on whether R could ship with more
> appropriate GC parameters.
I've been doing a number of similar measurements, and have come to the
same conclusion. R is currently very conservative about memory usage,
and this leads to unnecessarily poor performance on
2015 Jan 20
1
default min-v/nsize parameters
>>>>> Peter Haverty <haverty.peter at gene.com>
>>>>> on Mon, 19 Jan 2015 08:50:08 -0800 writes:
> Hi All, This is a very important issue. It would be very
> sad to leave most users unaware of a free speedup of this
> size. These options don't appear in the R --help
> output. They really should be added there.
Indeed,
1999 Oct 13
3
main/character.c (et.al): dangerous AllocBuffer()
I was hit by ugly crashes of R, when I tried to read big data sets ("volcano").
So I looked into the code and found the following in character.c (triggered by substr()):
I assume that the helper function AllocBuffer() shall facilitate an economic memory management. But
the use of realloc() in the else-branch does not conform to ANSI and may hit you with certain
compilers.
When called
1999 Oct 13
3
main/character.c (et.al): dangerous AllocBuffer()
I was hit by ugly crashes of R, when I tried to read big data sets ("volcano").
So I looked into the code and found the following in character.c (triggered by substr()):
I assume that the helper function AllocBuffer() shall facilitate an economic memory management. But
the use of realloc() in the else-branch does not conform to ANSI and may hit you with certain
compilers.
When called
2001 Aug 22
1
Huge workspace cannot be opened
Hi everyone,
I have a problem that some people may have already encountered but i did not
find the solution yet.
As I use R to simulate several arrays of data, my workspace is now 35Mb big and
I cannot launch R with it.
An "xdr real data read error occured" and R tells me to delete .RData or
increase memory. I WON'T delete this file and changing the max-nsize to 40600k
did not
2000 Sep 07
6
Writing a tabel in LaTeX-format
I have a vague recollection of seeing a R-function
that will output a table or matix in a format suitable
for the tabular enviroment in LaTeX.
But I cannot find it. Any hints?
Bendix
----------------------
Bendix Carstensen
Senior Statistician
Steno Diabetes Centre
Niels Steensens Vej 2
DK-2820 Gentofte
Denmark
tel: +45 44 43 87 38
mob: +45 28 25 87 38
fax: +45 44 43 73 13
bxc at novo.dk
2023 May 25
1
data.frame with a column containing an array
So is this an expected behavior or is it a bug which should be reported somewhere else?
Thanks!
Georg?
?
?
Gesendet:?Dienstag, 09. Mai 2023 um 19:28 Uhr
Von:?"Bert Gunter" <bgunter.4567 at gmail.com>
An:?"Georg Kindermann" <Georg.Kindermann at gmx.at>
Cc:?"Rui Barradas" <ruipbarradas at sapo.pt>, r-help at r-project.org
Betreff:?Re: [R] data.frame
2023 May 25
3
data.frame with a column containing an array
I really don't know. I would call it a request for extended capabilities of
[.data.frame, rather than a feature or bug. But maybe wiser heads than mine
who monitor this list can sort it out.
-- Bert
On Wed, May 24, 2023 at 8:52?PM Georg Kindermann <Georg.Kindermann at gmx.at>
wrote:
> So is this an expected behavior or is it a bug which should be reported
> somewhere else?
>
2015 Jan 17
0
default min-v/nsize parameters
Martin Morgan discussed this a year or so ago and as I recall bumped
up these values to the current defaults. I don't recall details about
why we didn't go higher -- maybe Martin does. I suspect the main
concern would be with small memory machines in student labs and less
developed countries. If there was a way on all platforms to identify
how much memory is available that might help to
2004 Oct 07
5
Question restricting ssh access for some users only
I've used ssh as a secure telnet up to now but done little else with
it. The FreeBSD machines I look after on our internet-facing network
all have one account which I connect to for administration. I've set
up /etc/hosts.allow on all the machines to only allow ssh from a
limited internal network range.
Now I want to create a new account on one machine which will be
accessible from the