Then you don't have enough memory to process the whole thing at once. Not
unlike stuffing your mouth with cookies and not being able to chew for lack of
space to move the food around in your mouth.
Now, can you answer my question?
On November 8, 2024 5:38:37 PM PST, Val <valkremk at gmail.com>
wrote:>The data was read. The problem is with processing.
>
>On Fri, Nov 8, 2024 at 7:30?PM Bert Gunter <bgunter.4567 at gmail.com>
wrote:
>>
>> Is the problem reading the file in or processing it after it has been
read in?
>>
>> Bert
>>
>> On Fri, Nov 8, 2024 at 5:13?PM Jeff Newmiller via R-help <r-help at
r-project.org> wrote:
>>>
>>> Can you tell us what is wrong with the "chunked" package
which comes up when you Google "r read large file in chunks"?
>>>
>>> On November 8, 2024 4:58:18 PM PST, Val <valkremk at
gmail.com> wrote:
>>> >Hi All,
>>> >
>>> >I am reading data file ( > 1B rows) and do some date
formatting like
>>> > dat=fread(mydatafile)
>>> > dat$date1 <- as.Date(ymd(dat$date1))
>>> >
>>> >However, I am getting an error message saying that
>>> > Error: cons memory exhausted (limit reached?)
>>> >
>>> >The script was working when the number rows were around
650M.
>>> >
>>> >Is there another way to handle a big data set in R?
>>> >
>>> >
>>> >Thank you.
>>> >
>>> >______________________________________________
>>> >R-help at r-project.org mailing list -- To UNSUBSCRIBE and
more, see
>>> >https://stat.ethz.ch/mailman/listinfo/r-help
>>> >PLEASE do read the posting guide
https://www.R-project.org/posting-guide.html
>>> >and provide commented, minimal, self-contained, reproducible
code.
>>>
>>> --
>>> Sent from my phone. Please excuse my brevity.
>>>
>>> ______________________________________________
>>> R-help at r-project.org mailing list -- To UNSUBSCRIBE and more,
see
>>> https://stat.ethz.ch/mailman/listinfo/r-help
>>> PLEASE do read the posting guide
https://www.R-project.org/posting-guide.html
>>> and provide commented, minimal, self-contained, reproducible code.
--
Sent from my phone. Please excuse my brevity.