Displaying 20 results from an estimated 500 matches similar to: "Parallel code using parLapply"
2013 Dec 24
2
Parallel computing: how to transmit multiple parameters to a function in parLapply?
Hi R-developers
In the package Parallel, the function parLapply(cl, x, f) seems to allow
transmission of only one parameter (x) to the function f. Hence in order to
compute f(x, y) parallelly, I had to define f(x, y) as f(x) and tried to
access y within the function, whereas y was defined outside of f(x).
Script:
library(parallel)
f <- function(x) {
z <- 2 * x + .GlobalEnv$y # Try to
2011 Feb 03
1
problem with parLapply from snow
Hi,
The following function use to work, but now it doesn't giving the error
"> CallSnow(, 100)
Using snow package, asking for 2 nodes
2 slaves are spawned successfully. 0 failed.
Error in checkForRemoteErrors(val) :
2 nodes produced errors; first error: no applicable method for 'lapply' applied to an object of class "list"
".
Where this is the
2005 Nov 11
1
Snow parLapply
Dear R-user,
I am trying to use the function 'parLapply' from the 'snow' package
which is supposed to work the same wys as 'lapply' but for a
parallelized cluster of computers. The function I am trying to call in
parallel is 'dudi.pca' (from the 'ade4' package) which performs
principal component analyses. When I call this function on a list of
2013 Feb 19
1
latin hypercube sampling
Hi all,
I am attempting to use latin hypercube sampling to sample different
variable functions in a series of simultaneous differential equations.
There is very little code online about lhs or clhs, so from different
other help threads I have seen, it seems I need to create a
probability density function for each variable function, and then use
latin hypercube sampling on this pdf.
So far, I
2017 Dec 11
0
document environment passing in parallel::parLapply
The runtime of parallel::parLapply depends on variables unrelated to
the parLapply call. However, this is not clearly documented. Therefore
I would like to suggest expanding the relevant documentation to
explain this behaviour.
Consider this example:
parallel_demo <- function(random_values_count) {
some_data <- runif(random_values_count)
dummy_function <- function(x) {
x
}
2012 Jun 20
2
Conditioned Latin Hypercube Sampling within determined distance
Hi all,
I am a begginer in R and I have been trying to use the Conditioned Latin
Hypercube to choose sample points only in areas close to roads due to the
difficult thorough access in the study area. I could use a code to create
the points throughout the area, but I need to create a code to make the R
only comes up with points close to this road. I've been using the packages
2011 Dec 02
2
Moving column averaging
# need zoo to use rollapply()
# your data (I called df)
df <- structure(list(a = 1:2, b = 2:3, c = c(5L, 9L), d = c(9L, 6L),
e = c(1L, 5L), f = c(4, 7)), .Names = c("a", "b", "c", "d",
"e", "f"), class = "data.frame", row.names = c(NA, -2L))
# transpose and make a zoo object
df2 <- zoo(t(df))
#rollapply to get
2014 Jul 02
1
parLapply on sqlQuery (from package RODBC)
R Version : 2.14.1 x64
Running on Windows 7
Connecting to a database on a remote Microsoft SQL Server 2012
The short form of my problem is the following.
I have an unordered vectors of names, say:
names<-c("A", "B", "A", "C","C")
each of which have an id in a table in my db. I need to convert the names to their corresponding ids.
I
2009 Nov 17
2
SVM Param Tuning with using SNOW package
Hello,
Is the first time I am using SNOW package and I am trying to tune the cost
parameter for a linear SVM, where the cost (variable cost1) takes 10 values
between 0.5 and 30.
I have a large dataset and a pc which is not very powerful, so I need to
tune the parameters using both CPUs of the pc.
Somehow I cannot manage to do it. It seems that both CPUs are fitting the
model for the same values
2008 Nov 30
2
Snow and multi-processing
Dear R gurus,
I have a very embarrassingly parallelizable job that I am trying to speed up with snow on our local cluster. Basically, I am doing ~50,000 t.test for a series of micro-array experiments, one gene at a time. Thus, I can easily spread the load across multiple processors and nodes.
So, I have a master list object that tells me what rows to pick up for each genes to do the t.test from
2018 Sep 12
1
Environments and parallel processing
On 12.09.2018 20:20, G?bor Cs?rdi wrote:
> This is all normal, a fork cluster works with processes, that do not
> share memory.
And if you are after shared-memory parallelism, you can try the 'Rdsm'
package: https://cran.r-project.org/package=Rdsm
Greetings
Ralf
--
Ralf Stubner
Senior Software Engineer / Trainer
daqana GmbH
Dortustra?e 48
14467 Potsdam
T: +49 331 23 61 93 11
F:
2018 Mar 14
2
clusterApply arguments
Hi!
I recognized that the argument matching of clusterApply (and therefore parLapply) goes wrong when one of the arguments of the function is called "c". In this case, the argument "c" is used as cluster and the functions give the following error message "Error in checkCluster(cl) : not a valid cluster".
Of course, "c" is for many reasons an unfortunate
2012 Aug 21
1
parLapply fails to detect default cluster?
invoking parLapply without a cluster fails to find a previously
registered cluster
> library(parallel)
> setDefaultCluster(makePSOCKcluster(2))
> parLapply(X=1:2, fun=function(...) {})
Error in cut.default(i, breaks) : invalid number of intervals
This is because in parLapply length(cl) is determined before
defaultCluster(cl) is called. By inspection, this appears to be true of
2012 Jan 12
1
parLapply within a function
Dear R users,
I have some problems with the parLapply function from the "parallel"
package:
I use parLapply on a pretty big R object without changing the object
within the called function. If I execute parLapply alone, everything
works fine. It seems that the object resides only once in the memory.
But if I use the same call within another function, the object seems to
be multiplied to
2007 Mar 27
2
snow parLapply standard output
I am slightly confused by the way the standard output is redirected in a R
snow cluster environment.
I am using parLapply from the snow package to execute a function on my
MPI/LAM cluster. How can I redirect standard output (produced using "cat")
from this function back to the terminal where I invoked it? I intend to
transmit some status information in advance to the final result of the
2010 Dec 02
1
parLapply - Error in do.call("fun", lapply(args, enquote)) : could not find function "fun"
Hello everybody,
I've got a bit of a problem with parLapply that's left me scratching my head
today. I've tried this in R 2.11 and the 23 bit Revolution R Enterprise and
gotten the same result, OS in question is Windows XP, the package involved
is the snow package.
I've got a list of 20 rain/no rain (1/0) situations for these two stations i
and j, all the items in this list look
2018 Mar 15
2
clusterApply arguments
Thank you for your answer!
I agree with you except for the 3 (Error) example and
I realize now I should have started with that in the explanation.
>From my point of view
parLapply(cl = clu, X = 1:2, fun = fun, c = 1)
shouldn't give an error.
This could be easily avoided by using all the argument
names in the custerApply call of parLapply which means changing,
parLapply <-
2020 Oct 29
2
Something is wrong with the unserialize function
Hi all,
I am not able to export an ALTREP object when `gctorture` is on in the
worker. The package simplemmap can be used to reproduce the problem. See
the example below
```
## Create a temporary file
filePath <- tempfile()
con <- file(filePath, "wrb")
writeBin(rep(0.0,10),con)
close(con)
library(simplemmap)
library(parallel)
cl <- makeCluster(1)
x <- mmap(filePath,
2018 Mar 15
1
clusterApply arguments
On 03/15/2018 05:25 PM, Henrik Bengtsson wrote:
> On Thu, Mar 15, 2018 at 3:39 AM, <FlorianSchwendinger at gmx.at> wrote:
>> Thank you for your answer!
>> I agree with you except for the 3 (Error) example and
>> I realize now I should have started with that in the explanation.
>>
>> From my point of view
>> parLapply(cl = clu, X = 1:2, fun = fun, c =
2018 Sep 12
2
Environments and parallel processing
While using parallelization R seems to clone all environments (that are normally passed by reference) that are returned from a child process. In particular, consider the following example:
library(parallel)
env1 <- new.env()
envs2 <- lapply(1:4, function(x) env1)
cl<-makeCluster(2, type="FORK")
envs3 <- parLapply(cl, 1:4, function(x) env1)
envs4 <- parLapply(cl, 1:4,