We do not know anything about the data, but given those are floting
point numbers or integers there sould not be a problem with 45000000
times 3.
Uwe Ligges
On 26.04.2012 05:15, Rich Lane wrote:> Hi all,
>
>
>
> I have an interesting project coming up, but the datasets are way bigger
> than anything I've used before with R. I'll end up with a dataset
with about
> 45,000,000 records, each with 3 columns. I'll probably want to add some
more
> columns for analysis if I can. My client can't deal with such big
files, so
> she is sending the data to me in chunks of "only" 4,500,000
records at a
> time. The question is, can I get away with making one dataset/matrix out of
> all 45 million records? It should be under 2 gigabytes total, but I
don't
> want the whole import process to fail after, say, 20 million records.
>
>
>
> I'm running Windows 7 Home Premium, 64 bit version, with 8 GB of RAM
and
> plenty of hard drive space, but I'm planning to upgrade all that just
for
> this project, so that it will complete in my lifetime.
>
>
>
> Thanks for any help!
>
>
>
> rich
>
>
> [[alternative HTML version deleted]]
>
> ______________________________________________
> R-help at r-project.org mailing list
> https://stat.ethz.ch/mailman/listinfo/r-help
> PLEASE do read the posting guide
http://www.R-project.org/posting-guide.html
> and provide commented, minimal, self-contained, reproducible code.