Displaying 20 results from an estimated 2000 matches similar to: "Parallel R"
2018 May 15
0
Systemfit
... and the mailing list is picky about attachments... whatever you attached did not conform to the stringent requirements mentioned in the Posting Guide. Pasting the code right into the email is usually safest, though you DO have to post using plain text (as the Posting Guide indicates) or your code may get mangled by the automatic html format removal.
On May 15, 2018 7:04:31 AM PDT, Bert Gunter
2008 Mar 10
1
crossprod is slower than t(AA)%*BB
Dear Rdevelopers
The background for this email is that I was helping a PhD student to
improve the speed of her R code. I suggested to replace calls like
t(AA)%*% BB by crossprod(AA,BB) since I expected this to be faster. The
surprising result to me was that this change actually made her code
slower.
> ## Examples :
>
> AA <- matrix(rnorm(3000*1000),3000,1000)
> BB <-
2018 May 15
1
Systemfit
Unless there is good reason not to, always cc the list -- there are lots of
smarter folks than I on it who can help.
I may or may not have time to look at this. Hopefully someone else will.
-- Bert
Bert Gunter
"The trouble with having an open mind is that people keep coming along and
sticking things into it."
-- Opus (aka Berkeley Breathed in his "Bloom County" comic strip
2008 Apr 17
1
Couldn't (and shouldn't) is.unsorted() be faster?
Hi,
Couldn't is.unsorted() bail out immediately here (after comparing
the first 2 elements):
> x <- 20000000:1
> system.time(is.unsorted(x), gcFirst=TRUE)
user system elapsed
0.084 0.040 0.124
> x <- 200000000:1
> system.time(is.unsorted(x), gcFirst=TRUE)
user system elapsed
0.772 0.440 1.214
Thanks!
H.
2018 May 15
2
Systemfit
OK, Let's try this again! Here is the reproducible script; it is long because I had to copy the panel dataset here. My question is related to systemfit; I don't know how to get the result for the entire panel.
#Reproducible script
Empdata<- read.csv("/Users/ngwinuiazenui/Documents/UPLOADemp.csv")
View(Empdata)
install.packages("systemfit")
2018 May 16
0
Systemfit
Sadly you failed to set your email program to send plain text and the data is corrupted at my end.
I also think you need to reduce the size of the data set... the intent here is to increase your understanding, not debug your particular analysis.
I will say that I am having a very challenging time understanding what you are trying to accomplish though. What are the equations that you think need
2008 Mar 10
2
write.table with row.names=FALSE unnecessarily slow?
write.table with large data frames takes quite a long time
> system.time({
+ write.table(df, '/tmp/dftest.txt', row.names=FALSE)
+ }, gcFirst=TRUE)
user system elapsed
97.302 1.532 98.837
A reason is because dimnames is always called, causing 'anonymous' row
names to be created as character vectors. Avoiding this in
src/library/utils, along the lines of
Index:
2006 May 14
1
Suggestion for system.time()
Hi, people. A tiny suggestion for the system.time function.
Could the returned vector have names? These could be like:
c("User", "System", "Elapsed", "Sub.User", "Sub.System")
That would then produce self-documenting output.
--
Fran?ois Pinard http://pinard.progiciels-bpi.ca
2008 Feb 16
2
R on a computer cluster
Dear all,
I usually run R on my laptop with Windows XP Professional.
Now I really want to run R on a computer cluster (4 processors) with
Suse Linux Enterprise ver. 10. But I am new with computer cluster.
Should I modify my functions in order to use the greater
performance
and availability than that provided by my laptop?
Is there any R
manual on parallel computations on
2013 Apr 26
1
Lubuntu 13.04 raring ringtail: Problems installing pnmath_0.0-4
Hola!
This is 64bit Lubuntu 13.04, with R-3.0.0 installed from
deb http://ppa.launchpad.net/marutter/rrutter/ubuntu raring main
via synaptic. gcc --version
gcc (Ubuntu/Linaro 4.7.3-1ubuntu1) 4.7.3
> sessionInfo()
R version 3.0.0 (2013-04-03)
Platform: x86_64-pc-linux-gnu (64-bit)
locale:
[1] LC_CTYPE=en_US.UTF-8 LC_NUMERIC=C
[3] LC_TIME=en_US.UTF-8 LC_COLLATE=en_US.UTF-8
2008 Jul 10
1
compiling pnmath on an intel processor running mac OS 10.5
Has anyone successfully compiled pnmath (http://www.stat.uiowa.edu/~luke/R/experimental
) for an intel processor running mac OS 10.5? When I attempt to do so
via the R package installer (choosing "Local Source Package" and
pointing to the pnmath_0.0-2.tar.gz file), I get the following errors:
* Installing *source* package 'pnmath' ...
** libs
** arch - i386
gcc -arch i386
2008 Feb 04
2
maybe a bug in the system.time() function? (PR#10696)
Full_Name: Alessandra Iacobucci
Version: 2.5.1
OS: Mac OS X 10.4.11
Submission from: (NULL) (193.48.71.92)
Hi,
I am making some intensive simulations for the testing of a Population Monte
Carlo algorithm. This involves also a study of the CPU times in two different
case.
What I am trying to measure is the "real" CPU time, the one which is independent
on the %CPU.
I'm using the
2008 Aug 21
1
pnmath compilation failure; dylib issue?
(1) ...need to speed up a monte-carlo sampling...any suggestions about
how I can get R to use all 8 cores of a mac pro would be most useful
and very appreciated...
(2) spent the last few hours trying to get pnmath to compile under os-
x 10.5.4...
using gcc version 4.2.1 (Apple Inc. build 5553) as downloaded from
CRAN, xcode 3.0...
...xcode 3.1 installed over top of above after
2010 Jun 04
5
R Newbie, please help!
Hello Everyone,
I just started a new job & it requires heavy use of R to analyze datasets.
I have a data.table that looks like this. It is sorted by ID & Date, there
are about 150 different IDs & the dataset spans 3 million rows. The main
columns of concern are ID, date, and totret. What I need to do is to derive
daily returns for each ID from totret, which is simply totret at time
2010 Nov 06
1
Hashing and environments
Hi,
I'm trying to write a general-purpose "lexicon" class and associated methods for storing and accessing information about large numbers of specific words (e.g., their frequencies in different genres). Crucial to making such a class practically useful is to get hashing working correctly so that information about specific words can be accessed quickly. But I've never really
2008 Jul 10
1
embarrassingly parallel problem - simple loop solution
I have an "embarrassingly parallel" routine that I need to run 24000^2/2
times (based on some microarray data). All I really need to do is
parallelize a nested for-loop. But I haven't found a clear list of what
packages/commands I'd need to do this. I've got a dual quad core xeon
system running RHEL5, so if I could use hyperthreading to increase the
number of (virtual)
2005 Aug 05
6
Computing sums of the columns of an array
Hi,
I have a 5x731 array A, and I want to compute the sums of the columns.
Currently I do:
apply(A, 2, sum)
But it turns out, this is slow: 70% of my CPU time is spent here, even
though there are many complicated steps in my computation.
Is there a faster way?
Thanks,
Martin
2005 Feb 25
3
Loops and dataframes
Hi,
I am experiencing a long delay when using dataframes inside loops and was
wordering if this is a bug or not.
Example code:
> st <- rep(1,100000)
> ed <- rep(2,100000)
> for(i in 1:length(st)) st[i] <- ed[i] # works fine
> df <- data.frame(start=st,end=ed)
> for(i in 1:dim(df)[1]) df[i,1] <- df[i,2] #takes for ever
R: R 2.0.0 (2004-10-04)
OS: Linux, Fedora Core 2
2004 Dec 06
6
how to get how many lines there are in a file.
hi all
If I wanna get the total number of lines in a big file without reading
the file's content into R as matrix or data frame, any methods or
functions?
thanks in advance.
Regards
2010 Aug 15
2
time of serialization
Hello,
I have question about the overhead in lapply.
x is a list of 3000 lists. Each of the i (1<=i<=3000) list elements is
pair of two elements: a string vector and a data frame
x is roughly 235MB.
> gc()
##
> z <- system.time(y <- lapply(x,function(r){
system.time(serialize(r,NULL))['elapsed']
}))
> sum(unlist(y))
18.812
> z
user system elapsed
494.144