Displaying 20 results from an estimated 700 matches similar to: "passing command line arguments to 'R CMD BATCH myScript.R'"
2010 Nov 02
2
get the scripname within the executed myscript.r
Dear R-users!
Is there a way to obtain the name of the executed myscript.r, i.e. when
cmd> rscript myscript.r
is executed? (The name of the script in this case is "myscript.r")
Here is the explanation of why I would like to get that: Why: I have
prepared a set of scripts (decompose_data.r, plot_with_higher_resolution.r,
plot_3D.r, etc.) placed in a root directory at one place. I
2024 Nov 09
3
Limit
Hi All,
I am reading data file ( > 1B rows) and do some date formatting like
dat=fread(mydatafile)
dat$date1 <- as.Date(ymd(dat$date1))
However, I am getting an error message saying that
Error: cons memory exhausted (limit reached?)
The script was working when the number rows were around 650M.
Is there another way to handle a big data set in R?
Thank you.
2005 Oct 04
2
newbie questions - looping through hierarchial datafille
Dear List,
Im new to R - making a transition from SAS. I have a space delimited file
with the following structure. Each line in the datafile is identified by
the first letter.
A = Inventory (Inventory)
X = Stratum (Stratum_no Total Ye=year established)
P = Plot (Plot_no age slope= species)
T = Tree (tree_no frequency)
L = Leader (leader diameter height)
F = Feature (start_height finish_height
2006 Nov 13
2
A printing "macro"
I am exploring the result of clustering a large multivariate data set
into a number of groups, represented, say, by a factor G.
I wrote a function to see how categorical variables vary between groups:
> ddisp <- function(dvar) {
+ csqt <- chisq.test(G,dvar)
+ print(csqt$statistic)
+ print(csqt$observed)
+ print(round(csqt$expected))
+ round(csqt$residuals)
+ }
>
> x
2024 Nov 09
1
Limit
Can you tell us what is wrong with the "chunked" package which comes up when you Google "r read large file in chunks"?
On November 8, 2024 4:58:18 PM PST, Val <valkremk at gmail.com> wrote:
>Hi All,
>
>I am reading data file ( > 1B rows) and do some date formatting like
> dat=fread(mydatafile)
> dat$date1 <- as.Date(ymd(dat$date1))
>
2024 Nov 09
1
Limit
Check the "high performance task view" on CRAN ...
https://cran.r-project.org/web/views/HighPerformanceComputing.html
On Fri, Nov 8, 2024, 7:58 PM Val <valkremk at gmail.com> wrote:
> Hi All,
>
> I am reading data file ( > 1B rows) and do some date formatting like
> dat=fread(mydatafile)
> dat$date1 <- as.Date(ymd(dat$date1))
>
> However, I
2010 Jun 06
2
Generalized DCC GARCH ML estimation
--
View this message in context: http://r.789695.n4.nabble.com/Generalized-DCC-GARCH-ML-estimation-tp2245125p2245125.html
Sent from the R help mailing list archive at Nabble.com.
2004 Aug 17
1
survdiff
Hello,
As I am quitte an ignorant user of R, excuse me for any wrongfull usage of
all the terms.
My question relates to the statistics behind the survdiff function in the
package survival.
My textbook knowledge of the logrank test tells me that if I want to compare
two survival curves, I have to take the sum of the factors: (O-E)^2/E of
both groups, which will give me the Chisq.
If I calculate
2024 Nov 09
1
Limit
Is the problem reading the file in or processing it after it has been read
in?
Bert
On Fri, Nov 8, 2024 at 5:13?PM Jeff Newmiller via R-help <
r-help at r-project.org> wrote:
> Can you tell us what is wrong with the "chunked" package which comes up
> when you Google "r read large file in chunks"?
>
> On November 8, 2024 4:58:18 PM PST, Val <valkremk at
2011 Feb 07
1
Question about checkTmvArgs function in rtmvnorm (package tmvtnorm)
Hello!
I was wondering if it's possible to see the actual code of
checkTmvArgs function that is part of the code for rtmvnorm (which is
below - I just typed "rtmvnorm" on the prompt). I get an error:
Error in checkTmvArgs(mean, sigma, lower, upper) :
sigma must be a symmetric matrix
At the same time I am pretty sure that the matrix I am passing as
sigma is a var-covar matrix
2009 Aug 19
4
Basic question: Reading in multiple choice question responses to a single column in data frame
I'm using read.delim to successfully read in tab delimited data, but some
columns' values are comma seperated, reflecting the fact that user chose a
few answers on a multi-select question. I understand that each answer is
its own category and so could be represented as a seperate column in the
data set, but I'd like the option of reading in the data column, and
converting it to a
2024 Nov 09
1
Limit
The data was read. The problem is with processing.
On Fri, Nov 8, 2024 at 7:30?PM Bert Gunter <bgunter.4567 at gmail.com> wrote:
>
> Is the problem reading the file in or processing it after it has been read in?
>
> Bert
>
> On Fri, Nov 8, 2024 at 5:13?PM Jeff Newmiller via R-help <r-help at r-project.org> wrote:
>>
>> Can you tell us what is wrong with
2024 Nov 09
1
Limit
Then you don't have enough memory to process the whole thing at once. Not unlike stuffing your mouth with cookies and not being able to chew for lack of space to move the food around in your mouth.
Now, can you answer my question?
On November 8, 2024 5:38:37 PM PST, Val <valkremk at gmail.com> wrote:
>The data was read. The problem is with processing.
>
>On Fri, Nov 8, 2024
2006 Mar 11
2
how to create analog stripchart plots of x vs t (t=mm/dd/yyyy hh:mm:ss)
Hello r-experts,
I sure could us a little help.
I have an ever updating text file with timestamped data in it. I can
reformat in anyway I want if need be but currently I have chosen to make
columns of date, time and measuresed value (comma delimeted and with the
dates and times in quotes to interpret them as strings).
Here is a small section of my text data file:
2009 Jun 25
2
stringsAsFactors has no impact in expand.grid()?
Hi
I have the feeling, that the argument stringsAsFactors has no impact in the
function expand.grid:
a <- c("PR", "NC", "A2", "BS")
b <- c(1, 0.5, 0.25, 0.125, 0.0625, 0.03125)
class(expand.grid(css, fscs, stringsAsFactors=FALSE)[[1]])
[1] "factor"
class(expand.grid(css, fscs, stringsAsFactors=TRUE)[[1]])
[1] "factor"
Also, when
2002 Jun 12
3
help debugging segfaults
(Sorry for the cross-post--- I wasn't sure which list is more
appropriate...)
Hi everyone,
I've run into segfaults when using my randomForest package on large dataset
(e.g., 100 x 15200) and large number of trees (e.g., ntree=7000 and
mtry=3000). I'm wondering if anyone can give me some hints on where to look
for the problem.
The randomForest package mainly consists of two things:
2002 Jun 12
3
help debugging segfaults
(Sorry for the cross-post--- I wasn't sure which list is more
appropriate...)
Hi everyone,
I've run into segfaults when using my randomForest package on large dataset
(e.g., 100 x 15200) and large number of trees (e.g., ntree=7000 and
mtry=3000). I'm wondering if anyone can give me some hints on where to look
for the problem.
The randomForest package mainly consists of two things:
2024 Nov 09
1
Limit
Hi Jeff,
Memory was not an issue. The system only used 75% of the memory
allocated for the job.
I am trying to understand what "r read large file in chunks" is doing.
On Fri, Nov 8, 2024 at 7:50?PM Jeff Newmiller <jdnewmil at dcn.davis.ca.us> wrote:
>
> Then you don't have enough memory to process the whole thing at once. Not unlike stuffing your mouth with cookies
2009 Feb 04
1
package ccgarch - dcc.estimation
Hello,
I am trying to model a bivariate time series called 'residuals' as a
dcc-garch model.
I want to use the function dcc.estimation(a, A, B dcc.para, dvar, model) to
estimate the parameters.
No matter how I tried to define a, A and B, I always got the message "Error
in constrOptim(theta = para, f = loglik.dcc2, gr = grad.dcc2, ui = resta, :
initial value not
2024 Nov 09
1
Limit
There is always an implied "and do computations on it before writing the processed data out" when reading chunks of a file.
And you would almost certainly not be getting that error if you were not out of memory. A good rule of thumb is that you need 4 times as much free memory to process data than you need to read it in.
On November 8, 2024 6:08:16 PM PST, Val <valkremk at