Michael Styer
2008-Nov-24 21:09 UTC
[R] increasing memory limit in Windows Server 2008 64-bit
Hello, I'm working with a very large dataset in R on a computer running 64-bit Windows Server 2008 Standard with 32GB of RAM. According to the R for Windows FAQ, the maximum value allowed for max-mem-size is 4095MB. Is it possible to run R with a higher memory limit on this system? I've tried changing memory.limit() in the R console but it claims the system has a 4-GB address limit, which I don't believe to be the case here. Thanks for your help. Mike
Michael Styer
2008-Nov-25 07:33 UTC
[R] Compiling R for 64-bit Windows (was Re: increasing memory limit in Windows Server 2008 64-bit)
I did some more research and I think I've answered my own question. So my next question is, does anyone have any thoughts about how significant a project it would be to compile R for 64-bit windows (using, e.g., the Portland Group compiler)? How much of the code would need to be changed to get around the issue with longs? I think this is probably beyond the scope of my project but I should make sure before I give up. Thanks for your help! Mike Michael Styer wrote:> Hello, > > I'm working with a very large dataset in R on a computer running 64-bit > Windows Server 2008 Standard with 32GB of RAM. According to the R for > Windows FAQ, the maximum value allowed for max-mem-size is 4095MB. Is it > possible to run R with a higher memory limit on this system? I've tried > changing memory.limit() in the R console but it claims the system has a > 4-GB address limit, which I don't believe to be the case here.
I'm currently working with very large datasets that consist out of 1,000,000 + rows. Is it at all possible to use R for datasets this size or should I rather consider C++/Java. -- View this message in context: http://www.nabble.com/increasing-memory-limit-in-Windows-Server-2008-64-bit-tp20675880p20699700.html Sent from the R help mailing list archive at Nabble.com.
Good afternoon, The short answer is "yes", the long answer is "it depends". It all depends on what you want to do with the data, I'm working with dataframes of a couple of million lines, on this plain desktop machine and for my purposes it works fine. I read in text files, manipulate them, convert them into dataframes, do some basic descriptive stats and tests on them, a couple of columns at a time, all quick and simple in R. There are some libraries which are setup to handle very large datasets, e.g. biglm [1]. If you're using algorithms which require vast quantities of memory, then as the previous emails in this thread suggest, you might need R running on 64-bit. If you're working with a problem which is "embarrassingly parallel"[2], then there are a variety of solutions - if you're in between then the solutions are much more data dependant. the flip question: how long would it take you to get up and running with the functionallity (tried and tested in R) you require if you're going to be re-working things in C++? I suggest that you have a look at R, possibly using a subset of your full set to start with - you'll be amazed how quickly you can get up and running. As suggested at the start of this email... "it depends"... Best Regards, Sean O'Riordain Dublin [1] http://cran.r-project.org/web/packages/biglm/index.html [2] http://en.wikipedia.org/wiki/Embarrassingly_parallel iwalters wrote:> > I'm currently working with very large datasets that consist out of > 1,000,000 + rows. Is it at all possible to use R for datasets this size > or should I rather consider C++/Java. > > >-- View this message in context: http://www.nabble.com/increasing-memory-limit-in-Windows-Server-2008-64-bit-tp20675880p20700590.html Sent from the R help mailing list archive at Nabble.com.