Displaying 20 results from an estimated 3000 matches similar to: "Problem reading RDS files"
2018 Apr 23
1
Problem reading RDS files
I've tried to re-experiment the tutorial presented at
http://www.rdatamining.com/docs/twitter-analysis-with-r and specifically
aimed to use rds files (tweet records) at http://www.rdatamining.com/data/.
On Sun, Apr 22, 2018 at 9:16 PM, Jeff Newmiller <jdnewmil at dcn.davis.ca.us>
wrote:
> Wouldn't the obvious problem be that your data file is corrupted or was
> never created
2018 Apr 22
0
Problem reading RDS files
Wouldn't the obvious problem be that your data file is corrupted or was never created using saveRDS in the first place? Can you show us a complete example of creating and attempting to read what was just created?
On April 22, 2018 10:20:05 AM CDT, mohammad moradi <mri.moradi at gmail.com> wrote:
>Hi there,
>
>I faced a weird problem doing a seemingly simple task in R.
2012 Sep 15
2
Risk of readRDS() not detecting race conditions with parallel saveRDS()?
I hardly know anything about the format used in (non-compressed)
serialization/RDS, but hoping someone with more knowledge could give
me some feedback;
Consider two R processes running in parallel on the same unknown file
system. Both of them write and read to the same RDS file foo.rds
(without compression) at random times using saveRDS(object,
file="foo.rds", compress=FALSE) and
2017 Aug 22
4
How to benchmark speed of load/readRDS correctly
Dear all
I was thinking about efficient reading data into R and tried several ways to test if load(file.Rdata) or readRDS(file.rds) is faster. The files file.Rdata and file.rds contain the same data, the first created with save(d, ' file.Rdata', compress=F) and the second with saveRDS(d, ' file.rds', compress=F).
First I used the function microbenchmark() and was a astonished
2017 Aug 22
0
How to benchmark speed of load/readRDS correctly
The large value for maximum time may be due to garbage collection, which
happens periodically. E.g., try the following, where the
unlist(as.list()) creates a lot of garbage. I get a very large time every
102 or 51 iterations and a moderately large time more often
mb <- microbenchmark::microbenchmark({ x <- as.list(sin(1:5e5)); x <-
unlist(x) / cos(1:5e5) ; sum(x) }, times=1000)
2018 Apr 03
1
[FORGED] recordPlot/replayPlot not working with saveRDS/readRDS
>>>>> Paul Murrell <paul at stat.auckland.ac.nz>
>>>>> on Tue, 3 Apr 2018 09:41:56 +1200 writes:
> Hi What you are doing "wrong" is loading a recordedplot
> into the same session that it was created in. The
> saveRDS()/readRDS() works if you save in one R session and
> then read in a different R session. The
2017 Aug 22
1
How to benchmark speed of load/readRDS correctly
Note that if you force a garbage collection each iteration the times are
more stable. However, on the average it is faster to let the garbage
collector decide when to leap into action.
mb_gc <- microbenchmark::microbenchmark(gc(), { x <- as.list(sin(1:5e5)); x
<- unlist(x) / cos(1:5e5) ; sum(x) }, times=1000,
control=list(order="inorder"))
with(mb_gc,
2018 Apr 02
3
recordPlot/replayPlot not working with saveRDS/readRDS
The documentation for recordPlot says the following:
> As of R 3.3.0, it is possible (again) to replay a plot from another R session using, for example, saveRDS and readRDS.
However, I haven't been able to save and restore a plot displaylist
and have it work within the same R session, using R 3.4.3 or 3.3.3.
Here's an example:
# Save displaylist for a simple plot
2012 Jul 29
1
readRDS, In as.double.xts(fishReport$count) : NAs introduced by coercion
Hello,
I looked in the R-help but could not find an archive addressing the
following. I would like to convert a character to numeric after reading a
file with RDS extension. After using as.numeric, I checked if it is
numeric. It was not converted. Please help.
Here is my code
>Report <- readRDS(file="RDS/Report.RDS")
> Report[1:2,]
dive_id date
2011 Oct 18
9
readRDS and saveRDS
Hi all,
Is there any chance that readRDS and saveRDS might one day become
read.rds and write.rds? That would make them more consistent with the
other reading and writing functions.
Hadley
--
Assistant Professor / Dobelman Family Junior Chair
Department of Statistics / Rice University
http://had.co.nz/
2018 Jan 27
0
Newbie wants to compare 2 huge RDSs row by row.
Hi Marsh,
An RDS is not a data structure such as a data.frame. It can be anything.
For example if I want to save my objects a, b, c I could do:
> saveRDS( list(a,b,c,), file="tmp.RDS")
Then read them back later with
> myList <- readRDS( "tmp.RDS" )
Do you have additional information about your "RDSs" ?
Eric
On Sat, Jan 27, 2018 at 6:54 AM, Marsh Hardy
2020 Jan 22
1
Memory error in the libcurl connection code
Hi All,
I think there is a memory error in the libcurl connection code that
typically happens when libcurl reads big chunks of data. This
potentially affects all code that use url() with the libcurl download
method, which is the default in most builds. In practice it tends to
happen more with HTTP/2 and if the connection is wrapped into a
gzcon(). macOS Catalina has a libcurl build with HTTP/2
2018 Jan 27
3
Newbie wants to compare 2 huge RDSs row by row.
Also, it will be easier to provide helpful information if you'd describe
what in your data you want to compare and what you hope to get out of the
comparison.
Best wishes,
Ulrik
Eric Berger <ericjberger at gmail.com> schrieb am Sa., 27. Jan. 2018, 08:18:
> Hi Marsh,
> An RDS is not a data structure such as a data.frame. It can be anything.
> For example if I want to save my
2012 Jan 09
1
serializing recordedplot object
I use recordPlot() to save plots to disk that I render later to a
variety of formats. This works fine for base R plots and ggplot2
plots, and also used to work for lattice plots. However somewhere in
version 2.14 things stopped working for lattice plots. Here is an
example:
library(lattice);
histogram(rnorm(100));
x <- recordPlot();
saveRDS(x, "myplot.rds");
y <-
2018 Jan 27
0
Newbie wants to compare 2 huge RDSs row by row.
Hi Guys, I apologize for my rank & utter newness at R.
I used summary() and found about 95 variables, both character and numeric, all with "Length:368842" I assume is the # of records.
I'd like to know the record number (row #?) of any record where the data doesn't match in the 2 files of what should be the same output.
Thanks in advance, M.
//
2016 Jul 20
4
un solo un favor
Hola a todos
Esta es mi primera pregunta en el grupo, y es sencilla pero me tiene
atascado. Estoy tratando de cortar mi mapa de (poner limites en UTM) en un
lugar definido como mi area de estudio (en este caso el sur de chile). Pero
creo no estar usando bien la funciĆ³n CRS ponendo bien los limites
requeridos.
> study_area <- readRDS("CHL_adm0.rds")
> study_area_UTM <-
2018 Mar 22
3
exporting data to stata
Hi,
I am new to R and I want to export data into Stata. Could somebody help with that? Thanks a lot.
This is the code I am using:
> setwd("D:/datasets/Seg-bcn/ESBD")
> data1 <- readRDS("r17045_ESDB_Habitatges_BDD_V_1_0.rds")
> library(foreign)
> write.dta(data="data1", file = "D:/datasets/data1.dta")
Error in write.dta(data =
2018 Apr 20
1
Problem reading R files
Dear R Developers,
I hope this email finds you well.
As a newcomer to the world of R, I'm trying different exercises. In my way,
I faced with some, seemingly, trivial problems but there is not any
response for that around the web. So, I thought can get help from you.
Specifically, in tutorial presented at http://www.rdatamining.com/doc
s/twitter-analysis-with-r, I faced with a problem on
2018 Jan 27
2
Newbie wants to compare 2 huge RDSs row by row.
If your two objects have class "data.frame" (look at class(objectName)) and
they
both have the same number of columns and the same order of columns and the
column types match closely enough (use all.equal(x1, x2) for that), then
you can try
which( rowSums( x1 != x2 ) > 0)
E.g.,
> x1 <- data.frame(X=1:5, Y=rep(c("A","B"),c(3,2)))
> x2 <-
2018 Jan 27
2
Newbie wants to compare 2 huge RDSs row by row.
Each RDS is 40 MBs. What's a slick code to compare them row by row, IDing row numbers with mismatches?
Thanks in advance.
//