similar to: R for large data

Displaying 20 results from an estimated 9000 matches similar to: "R for large data"

2001 Jul 24
3
Memory/data -last time I promise
I've seen several posts over the past 2-3 weeks about memory issues. I've tried to carefully follow the suggestions, but remain baffled as to why I can't load data into R. I hope that in revisiting this issue that I don't exasperate the list. The setting: 1 gig RAM , Linux machine 10 Stata files of approximately 14megs each File contents appear at the end of this boorishly long
2001 Jul 24
3
Memory/data -last time I promise
I've seen several posts over the past 2-3 weeks about memory issues. I've tried to carefully follow the suggestions, but remain baffled as to why I can't load data into R. I hope that in revisiting this issue that I don't exasperate the list. The setting: 1 gig RAM , Linux machine 10 Stata files of approximately 14megs each File contents appear at the end of this boorishly long
2015 Feb 15
4
Centos 7.0 and mismatched swap file
Everyone, I am putting together a new mail server for our firm using a SuperMicro with Centos 7.0. When performed the install of the os, I put 16 gigs of memory in the wrong slots on the mother board which caused the SuperMicro to recognize 8 gigs instead of 16 gigs. When I installed Centos 7.0, this error made the swap file 8070 megs instead of what I would have expected to be a over 16000
2001 Jul 19
2
Append/merge
I have a question about stacking datasets. I have 40 stata datasets that have exactly the same number of variables, with the same names (~420k rows, 8 columns). The datasets are relatively large ~ 15 megs. If they were text files a linux "cat file1 file2 >> combo" sort of strategy would work. I've considered using a merge command, but I don't want any records merged,
2007 Sep 22
2
install freezes at partition layout
I'm trying to install centos 5 on an old machine, a Micron with a pentium III and 250 megs of ram. I'm using the DVD image. The media check said the DVD was OK. Everything starts fine (both in graphics and text modes), untl I get to the custom partition table. The new system will go on hda, with the boot and root partition reformatted (see below). I laid it out to preserve the data on
2008 Jul 11
2
some files could not be transferred (code 23) at main.c (1506)
Hi, I'm trying to rsync a fairly large chunk of data [around 350 gigs, mostly 1 gig files] between 2 hosts in separate colo facilities. Link is plenty fast, I can pull 10 megs easily. Everytime, it seems to die with something like this, whether I try rsync from the stable 2 branch or the beta 3 branch sent 453675 bytes received 956704928 bytes 404632.68 bytes/sec total size is
2004 May 04
3
increasing memory
Hi. I want to use R with very large files, a couple hundred megabytes, but R crashes every time that I try. Reading help files seems to indicate that R ought to manage its memory itself. I know I have enough memory since stata handles these files perfectly well. I have a mac running OS 10.3 and am running RAqua 1.8.1. Is there anything I can do to make it deal with these files successfully?
2013 Aug 27
6
Suggest changing dirhash defaults for FreeBSD 9.2.
I have been experimenting with dirhash settings, and have scoured the internet for other peoples' experience with it. (I found the performance improvement in compiling has forestalled the need to add an SSD drive. ;-) I believe that increasing the following values by 10 would benefit most FreeBSD users without disadvantage. vfs.ufs.dirhash_maxmem: 2097152 to 20971520
2012 Nov 13
5
Getting information encoded in a SAS, SPSS or Stata command file into R.
Dear folks ? I have a large (26 gig) ASCII flat file in fixed-width format with about 10 million observations of roughly 400 variables. (It is 51 years of Current Population Survey micro data from IPUMS, roughly half the fields for each record). The file was produced by automatic process in response to a data request of mine. The file is not accompanied by a human-readable file giving the
2011 Sep 07
1
randomForest memory footprint
Hello, I am attempting to train a random forest model using the randomForest package on 500,000 rows and 8 columns (7 predictors, 1 response). The data set is the first block of data from the UCI Machine Learning Repo dataset "Record Linkage Comparison Patterns" with the slight modification that I dropped two columns with lots of NA's and I used knn imputation to fill in other gaps.
2001 Nov 02
2
no free space on root?
I recently upgraded from 7.1 to 7.2 and told it to upgrade my root filesystem (5gig) to Ext3. A while after that, I ran out of space. Directly after upgrading I had approximately 1.2gig free space. Thinking that I needed more space, I proceeded to create 3 extended partitions for /home, /usr and /tmp. Mounted them in /mnt, copied files..etc. All of these filesystems were left as ext2.
2007 Jan 17
4
Memory leak with character arrays?
Hi - When I'm trying to read in a text file into a labeled character array, the memory stamp/footprint of R will exceed 4 gigs or more. I've seen this behavior on Mac OS X, Linux for AMD_64 and X86_64., and the R versions are 2.4, 2.4 and 2.2, respectively. So, it would seem that this is platform and R version independant. The file that I'm reading contains the upstream regions
2014 Jan 23
2
Stata support in package foreign
As you know, Stata support in 'foreign' was frozen a while back at Stata version 12. R-core has received a request from a grad student to 'give top priority' to supporting Stata 13. That is not going to happen, not least because none of us has access to Stata. However, according to Stata's documentation both Stata 12 and 13 use format 115, so foreign::read.dta should be
2001 Nov 14
1
Embedded Cross-Compiling of OpenSSL & OpenSSH
I'm developing a minimal embedded linux system that will be running on a powerpc chip. I can get everything to work but the size is quite large. The majority of my size requirements are in libcrypto.a which is around 1.52 Megs I tried to pare it down further by removing some crypto options [ Configure linux-ppc no-asm no-cast no-des no-dh no-md2 no-mdc2 no-rc2 no-rc4 no-rc5 shared ] This
2004 Apr 29
3
Dummies in R
Dear all, my problem is following. I know Stata, but currently I have to use R. Could You please help in finding the analogy to R. (1) creating of City-Dummy. I know in Stata: g byte city=0 replace city=1 if city==12&year==2000 and (2) Create a Time-Dummy-Variable g byte T2000=0 replace T2000=1 if year==2000 (3) I need the City DUmmy for the following combination: I have the
2009 Jun 29
2
Large Stata file Import in R
Hi I am using Stata 10 and I need to import a data set in stata 10 to R, I have saved the dataset in lower versions of Stata as well by using saveold command in Stata. My RAM is 4gb and the stata file is 600MB, I am getting an error message which says : "Error: cannot allocate vector of size 3.4 Mb In addition: There were 50 or more warnings (use warnings() to see the first 50)" Thus
2008 Aug 24
3
Virtual Disk Size for C drive in wine
I start to install my program and I get an error that I have no space available.. 1 gig free on my Puppy system.. and the program only needs 75 megs... how can I change the settings in wine to show more megs. wine 0.9.22
2009 Dec 16
1
Stata files
I have a client who uses Stata 11. Can anyone advise me on ways of transferring data from this version of Stata to R? Reading the documentation for read.dta in package foreign it seems it only allows for versions up to Stata 10. I don't know Stata at all but a bit of poking around on their website led me to believe Stata 11 couldn't write in Stata 10 format. However Stata 11 can
2015 Feb 16
0
Centos 7.0 and mismatched swap file
On 2015-02-15, Gregory P. Ennis <PoMec at PoMec.Net> wrote: > > I am putting together a new mail server for our firm using a SuperMicro > with Centos 7.0. When performed the install of the os, I put 16 gigs of > memory in the wrong slots on the mother board which caused the > SuperMicro to recognize 8 gigs instead of 16 gigs. When I installed > Centos 7.0, this error made
2015 Feb 16
1
Centos 7.0 and mismatched swap file
On Sun, 2015-02-15 at 17:37 -0800, Keith Keller wrote: > On 2015-02-15, Gregory P. Ennis <PoMec at PoMec.Net> wrote: > > > > I am putting together a new mail server for our firm using a SuperMicro > > with Centos 7.0. When performed the install of the os, I put 16 gigs of > > memory in the wrong slots on the mother board which caused the > > SuperMicro to