Displaying 20 results from an estimated 5000 matches similar to: "R on os390"
2000 Aug 16
0
R on os390
G'day R friends,
I was wondering if anyone's ported R to os390. If so, are the vsize and
nsize limits the same as other platforms?
I could really annoy those SAS guys then.
thanks,
John Strumila
john.strumila at team.telstra.com
-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-
r-help mailing list -- Read http://www.ci.tuwien.ac.at/~hornik/R/R-FAQ.html
2001 Jan 04
6
regression constraints?
gday R gurus,
I have a multivariate regression for which I want to constrain the
coefficients to be > 0. Is this possible?
I've check the doco and searched CRAN but can't find anything.
thanks,
John Strumila
-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-
r-help mailing list -- Read http://www.ci.tuwien.ac.at/~hornik/R/R-FAQ.html
Send
2000 Sep 04
3
somebody's stolen my probabilities
gday gurus,
could someone please help me with this one? Can't find it in the doco.
summary of my lm object gives "Pr(>|t|)" but coef doesn't. How do I get
hold of these probabilities within an object? There doesn't seem to be an
appropriate attribute within the object - I was hoping for something like
"my.lm$probabilies".
I was hoping to grab the best t
2003 Sep 19
2
Latest Samba Disto for Os390/Unix ?
HI,I would be interested if there is a release of Samba for Os390/Unix
later that 1.9.
(I have tried downloading the 'samba-latest.tar.gz' file from the mirror
site,but as I cannot unzip/untar it,due to not having the 'gzip' program
!).
Therfore I am not sure how compatible it may or may not be with my
Os390/Uss platform.
Thanks for any help received in advance.
John.
2001 Feb 19
1
ghostscript errors
gday R friends,
Since I had random errors with ghostscript 6.01 I thought I'd upgrade.
After upgrading to gs6.5 I've had problems. Originally, gs failed
complaining about lack of "__sysconf" in glibc2.2 so I upgraded glibc. Now I
get this. Note that it used to work with gs6.01.
I realise this is probably a ghostscript problem but you never know...
thanks,
John Strumila
>
2001 Jan 11
1
lm variable name limit?
gday R gurus,
has anyone seen this lm problem?
When I add 2 bytes to the variable name (via names), lm returns an error.
It's as if there's a limit to the size of the variable names.
Any suggestions?
thanks,
John Strumila
> names(y)[12:14]
[1] "Log.SQLServer.Locks.RID.Lock.Wait" "Log.SQLServer.Lock"
[3] "xxaa"
2000 Mar 03
1
tapply, sorting and the heap
howdy gurus,
I'm new and green and I was hoping for a tiny bit of your expertise.
I'm running out of virtual memory (heap?) when summing using tapply. I've
already used --vsize=90M on my hpux machine. (details below)
Can I pre-sort or something to prevent my error?
thanks,
John Strumila
john.strumila at corpmail.telstra.com.au
> gc()["Vcells","total"]
[1]
2000 May 30
6
heap size trouble
Hi ,
I ''ve got a trouble with using R.
When I want to load a file that contains 93 thousand raws and 22 colums
of data (essentially float)
R shows me this error message
"heap size trouble"
Does anyone could tell me what parameter shall I precise before
launching R in order to load my big file.
Thanks a lot
-------------- next part --------------
A non-text attachment was
2002 Jan 07
1
internet2 and proxies
gday R gurus,
I've never attempted to use download.file before on my win2k machine but
this version 1.4 looks exciting. However, I'm not getting something right.
I've passed my R shortcut --internet2 and I can tell it's using the wininet
calls because it's asking about authentication but how do I tell it my
username/password?
The help file only talks about user/pw for non
2001 May 01
3
Factor Analysis
Thanks to Brian Ripley for the time and effort put into developing the
factanal package for R. I have found it very useful. However, in trying to
replicate some results from previous research, I have run into the need for
a couple of extensions and was wondering if they might find their way into
future implementations.
1. Though maximum likelihood estimates might be more rigorous and
2000 Nov 09
3
maximum of nsize=20000k ??
Dear R-ers,
somehow it is not possible to increase nsize to more than
20000k. When I specify e.g.
> R --vsize=10M --nsize=21000K
the result is:
free total (Mb)
Ncells 99658 350000 6.7
Vcells 1219173 1310720 10.0
Maybe I have overlooked s.th....
Marcus
--
+-------------------------------------------------------
| Marcus Eger
| E-Mail: eger.m at gmx.de (NEW)
|
2000 Oct 04
5
postscript preview
gday R friends,
I was wondering if anyone could help me with some graphics issues please.
I'm processing R in batch on redhat6.2 with some graphics to be imported
into msword on NT. For scaling within my doco I want some vector type
graphics. win.metafile doesn't seem to be supported on redhat which gives
me postscript. postscript is ideal but doesn't contain a preview so I
1999 Nov 12
1
R-0.65.1 Startup
Dear R users,
I have noticed that my R startup is extremely slow. It takes almost 3
minutes from "double-click" to R prompt. I have been running R-0.64.1 till
recently and it took about 30 sec. I still have access to R-0.64.1. When I
started it up, it took about 25 sec. Can anyone tell me if this is a bug in
R or a problem with my machine?
Note: This is after bootup with R being the
2005 Jun 29
3
Memory Management under Linux: Problems to allocate large amounts of data
Dear Group
I'm still trying to bring many data into R (see older postings). After solving some troubles with the database I do most of the work in MySQL. But still I could be nice to work on some data using R. Therefore I can use a dedicated Server with Gentoo Linux as OS hosting only R. This Server is a nice machine with two CPU and 4GB RAM which should do the job:
Dual Intel XEON 3.06 GHz
2010 Jan 19
2
Server hanging despite efforts to correct memory limits
My group is working with datasets between 100 Mb and 1 GB in size, using
multiple log ins. From the documentation, it appears that vsize is limited
to 2^30-1, which tends to prove too restrictive for our use. When we drop
that restriction (set vsize = NA) we end up hanging the server, which
requires a restart. Is there any way to increase the memory limits on R
while keeping our jobs from
1999 May 15
2
vsize and nsize
I am running R version ??? under Redhat 5.2. It seems as though the
--nsize object has no effct on the size of the allocated Ncells as
determined using gc(). Yes, I have that much data....
That is if I envoke R with
R --vsize 100 --nsize 5000000
then type
gc()
I get
free total
Ncells 92202 200000
Vcells 12928414 13107200
Thanks
Tony Long
Ecology and Evolutionary Biology
Steinhaus
1999 Apr 27
2
Memory management
Dear all,
I don't get it:
First of all, the help doesn't say what are the memory limits of
R. Say, what's the max heap size for instance ????
Secondly, I invoke R with the following commands each time:
rgui --vsize 30M --nsize 1000K
rgui --vsize 30M --nsize 2000K
rgui --vsize 30M --nsize 3000K
rgui --vsize 30M --nsize 4000K
I try to open a matrix 8000x8000 by issuing
2000 Oct 02
3
R vs S-PLUS with regard to memory usage
I am trying to translate code from S-PLUS to R and R really struggles!
After starting R with the foll.
R --vsize 50M --nsize 6M --no-restore
on a 400 MHz Pentium with 192 MB of memory running Linux (RH 6.2),
I run a function that essentially picks up an external dataset with 2121
rows
and 30 columns and builds a lm() object and also runs step() ... the step()
takes forever to run...(takes very
1999 Apr 12
3
--nsize and --vsize
Martin M has suggested I widen this discussion to R-devel, and
> I agree that we should increase them,
> but I'm not sure at all about the amount.
>
> The default could even depend on the architecture (via "./configure")..
Views, please.
------------- Begin Forwarded Message -------------
Is is not time we increased the defaults a bit? As the base gets bigger
I hit
2009 May 07
1
increasing memory for R bg job
Hi,
Is the following command used to increase the memory or any other command when a background R job is run?
R --min-vsize=vl --max-vsize=vu --min-nsize=nl --max-nsize=nu --max-ppsize=N
source:
http://stat.ethz.ch/R-manual/R-patched/library/base/html/Memory.html
Thx
Carol
[[alternative HTML version deleted]]