Displaying 20 results from an estimated 9000 matches similar to: "R for large data"
2001 Jul 24
3
Memory/data -last time I promise
I've seen several posts over the past 2-3 weeks about memory issues. I've
tried to carefully follow the suggestions, but remain baffled as to why I
can't load data into R. I hope that in revisiting this issue that I don't
exasperate the list.
The setting:
1 gig RAM , Linux machine
10 Stata files of approximately 14megs each
File contents appear at the end of this boorishly long
2001 Jul 24
3
Memory/data -last time I promise
I've seen several posts over the past 2-3 weeks about memory issues. I've
tried to carefully follow the suggestions, but remain baffled as to why I
can't load data into R. I hope that in revisiting this issue that I don't
exasperate the list.
The setting:
1 gig RAM , Linux machine
10 Stata files of approximately 14megs each
File contents appear at the end of this boorishly long
2015 Feb 15
4
Centos 7.0 and mismatched swap file
Everyone,
I am putting together a new mail server for our firm using a SuperMicro
with Centos 7.0. When performed the install of the os, I put 16 gigs of
memory in the wrong slots on the mother board which caused the
SuperMicro to recognize 8 gigs instead of 16 gigs. When I installed
Centos 7.0, this error made the swap file 8070 megs instead of what I
would have expected to be a over 16000
2001 Jul 19
2
Append/merge
I have a question about stacking datasets.
I have 40 stata datasets that have exactly the same number of variables,
with the same names (~420k rows, 8 columns).
The datasets are relatively large ~ 15 megs.
If they were text files a linux "cat file1 file2 >> combo" sort of
strategy would work.
I've considered using a merge command, but I don't want any records
merged,
2007 Sep 22
2
install freezes at partition layout
I'm trying to install centos 5 on an old machine, a
Micron with a pentium III and 250 megs of ram. I'm
using the DVD image.
The media check said the DVD was OK.
Everything starts fine (both in graphics and text
modes), untl I get to the custom partition table. The
new system will go on hda, with the boot and root
partition reformatted (see below). I laid it out to
preserve the data on
2008 Jul 11
2
some files could not be transferred (code 23) at main.c (1506)
Hi,
I'm trying to rsync a fairly large chunk of data [around 350 gigs,
mostly 1 gig files] between 2 hosts in separate colo facilities. Link
is plenty fast, I can pull 10 megs easily.
Everytime, it seems to die with something like this, whether I try
rsync from the stable 2 branch or the beta 3 branch
sent 453675 bytes received 956704928 bytes 404632.68 bytes/sec
total size is
2004 May 04
3
increasing memory
Hi. I want to use R with very large files, a couple hundred megabytes,
but R crashes every time that I try.
Reading help files seems to indicate that R ought to manage its memory
itself. I know I have enough memory since stata handles these files
perfectly well. I have a mac running OS 10.3 and am running RAqua 1.8.1.
Is there anything I can do to make it deal with these files successfully?
2013 Aug 27
6
Suggest changing dirhash defaults for FreeBSD 9.2.
I have been experimenting with dirhash settings, and have scoured the internet
for other peoples' experience with it.
(I found the performance improvement in compiling has forestalled the need to
add an SSD drive. ;-)
I believe that increasing the following values by 10 would benefit most FreeBSD
users without disadvantage.
vfs.ufs.dirhash_maxmem: 2097152 to 20971520
2012 Nov 13
5
Getting information encoded in a SAS, SPSS or Stata command file into R.
Dear folks ?
I have a large (26 gig) ASCII flat file in fixed-width format with about 10
million observations of roughly 400 variables. (It is 51 years of Current
Population Survey micro data from IPUMS, roughly half the fields for each
record). The file was produced by automatic process in response to a data
request of mine.
The file is not accompanied by a human-readable file giving the
2011 Sep 07
1
randomForest memory footprint
Hello, I am attempting to train a random forest model using the
randomForest package on 500,000 rows and 8 columns (7 predictors, 1
response). The data set is the first block of data from the UCI
Machine Learning Repo dataset "Record Linkage Comparison Patterns"
with the slight modification that I dropped two columns with lots of
NA's and I used knn imputation to fill in other gaps.
2001 Nov 02
2
no free space on root?
I recently upgraded from 7.1 to 7.2 and told it to upgrade my root
filesystem (5gig) to Ext3.
A while after that, I ran out of space. Directly after upgrading I had
approximately 1.2gig free space. Thinking that I needed more space, I
proceeded to create 3 extended partitions for /home, /usr and /tmp. Mounted
them in /mnt, copied files..etc. All of these filesystems were left as
ext2.
2007 Jan 17
4
Memory leak with character arrays?
Hi -
When I'm trying to read in a text file into a labeled character array,
the memory stamp/footprint of R will exceed 4 gigs or more. I've seen
this behavior on Mac OS X, Linux for AMD_64 and X86_64., and the R
versions are 2.4, 2.4 and 2.2, respectively. So, it would seem that
this is platform and R version independant.
The file that I'm reading contains the upstream regions
2014 Jan 23
2
Stata support in package foreign
As you know, Stata support in 'foreign' was frozen a while back at Stata
version 12. R-core has received a request from a grad student to 'give
top priority' to supporting Stata 13. That is not going to happen, not
least because none of us has access to Stata.
However, according to Stata's documentation both Stata 12 and 13 use
format 115, so foreign::read.dta should be
2001 Nov 14
1
Embedded Cross-Compiling of OpenSSL & OpenSSH
I'm developing a minimal embedded linux system that will be running on a
powerpc chip. I can get everything to work but the size is quite large.
The majority of my size requirements are in libcrypto.a which is around
1.52 Megs
I tried to pare it down further by removing some crypto options
[ Configure linux-ppc no-asm no-cast no-des no-dh no-md2 no-mdc2 no-rc2
no-rc4 no-rc5 shared ]
This
2004 Apr 29
3
Dummies in R
Dear all,
my problem is following. I know Stata, but currently I have to use R. Could You please help in finding the analogy to R.
(1) creating of City-Dummy.
I know in Stata:
g byte city=0
replace city=1 if city==12&year==2000
and
(2) Create a Time-Dummy-Variable
g byte T2000=0
replace T2000=1 if year==2000
(3) I need the City DUmmy for the following combination: I have the
2009 Jun 29
2
Large Stata file Import in R
Hi
I am using Stata 10 and I need to import a data set in stata 10 to R, I have
saved the dataset in lower versions of Stata as well by using saveold
command in Stata.
My RAM is 4gb and the stata file is 600MB, I am getting an error message
which says :
"Error: cannot allocate vector of size 3.4 Mb
In addition: There were 50 or more warnings (use warnings() to see the first
50)"
Thus
2008 Aug 24
3
Virtual Disk Size for C drive in wine
I start to install my program and I get an error that I have no space available.. 1 gig free on my Puppy system.. and the program only needs 75 megs... how can I change the settings in wine to show more megs. wine 0.9.22
2009 Dec 16
1
Stata files
I have a client who uses Stata 11.
Can anyone advise me on ways of transferring data from this version of
Stata to R?
Reading the documentation for read.dta in package foreign it seems it
only allows for versions up to Stata 10. I don't know Stata at all but a
bit of poking around on their website led me to believe Stata 11
couldn't write in Stata 10 format. However Stata 11 can
2015 Feb 16
0
Centos 7.0 and mismatched swap file
On 2015-02-15, Gregory P. Ennis <PoMec at PoMec.Net> wrote:
>
> I am putting together a new mail server for our firm using a SuperMicro
> with Centos 7.0. When performed the install of the os, I put 16 gigs of
> memory in the wrong slots on the mother board which caused the
> SuperMicro to recognize 8 gigs instead of 16 gigs. When I installed
> Centos 7.0, this error made
2015 Feb 16
1
Centos 7.0 and mismatched swap file
On Sun, 2015-02-15 at 17:37 -0800, Keith Keller wrote:
> On 2015-02-15, Gregory P. Ennis <PoMec at PoMec.Net> wrote:
> >
> > I am putting together a new mail server for our firm using a SuperMicro
> > with Centos 7.0. When performed the install of the os, I put 16 gigs of
> > memory in the wrong slots on the mother board which caused the
> > SuperMicro to