similar to: Syntax for Rmpi cf multicore

Displaying 20 results from an estimated 300 matches similar to: "Syntax for Rmpi cf multicore"

2010 Jul 12
1
How to use mpi.allreduce() in Rmpi?
Hi everybody! I have the next code which makes a reduction of the *a *variable in two slaves, using the Rmpi package. library(Rmpi) mpi.spawn.Rslaves(nslaves=2) reduc<-function(){ a<-mpi.comm.rank()+2 mpi.reduce(a,type=2, op="prod") return(paste("a=",a)) } mpi.bcast.Robj2slave(reduc) mpi.remote.exec(reduc()) cat("Product: ")
2010 Jul 12
1
How to use mpi.allreduce() in Rmpi?
Hi everybody! I have the next code which makes a reduction of the *a *variable in two slaves, using the Rmpi package. library(Rmpi) mpi.spawn.Rslaves(nslaves=2) reduc<-function(){ a<-mpi.comm.rank()+2 mpi.reduce(a,type=2, op="prod") return(paste("a=",a)) } mpi.bcast.Robj2slave(reduc) mpi.remote.exec(reduc()) cat("Product: ")
2008 Nov 07
1
Rmpi task-pull
Hi, I'm testing the efficiency of the Rmpi package regarding parallelization using a cluster. I've found and tried the task pull programming method, but even if it is described as the best method, it seems to cause deadlock, anyone could help me in using this method? here is the code I've found and tried: # Initialize MPI library("Rmpi") # Notice we just say "give us
2009 Mar 25
0
Rmpi - send/receive multiple objects to slaves
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 I've written a function that uses Rmpi to perform a calculation in parallel. It works fine, but I'm trying to improve efficiency in terms of memory usage and the amount of data being passed back and forth between mater and slaves. Calculations are performed on a symmetrical matrix in order to zero-out some of the cells. In the parallel
2007 Jul 05
0
Question on Rmpi looping
Dear R list, In the course of learning to work with Rmpi, we are confused about a few points. The following simple program is based on some examples we retrieved from the web. Each slave is writing the same output line multiple times (a multiple equal to the number of slaves). In other words, the write statements are being executed a number of times equal to the number of slaves. I am
2008 Jul 16
1
Problem with mpi.close.Rslaves()
I am running R 2.7.0 on a Suse 9.1 linux cluster with a job scheduler dispatching jobs and openmpi-1.0.1. I have tried running one of the examples at http://ace.acadiau.ca/math/ACMMaC/Rmpi/examples.html in Rmpi and they seem to be working, except mpi.close.Rslaves() hangs. The slaves are closed, but the master doesn't finish its script. Below is the example script and the call to R. The job is
2010 Aug 12
0
send and receiving objects with Rmpi
Hi all, I'm having difficulties passing data from the master to the children. I haven't seen any complete tutorial, all the tutorial broadcast data where I would like to send parts of my big tablechip list. Everytime I run this, the last line always jams. I even tried to reduce the number of rows in data to 10 and still doesn't work. I even tried to do a remote.exec(ls()) and
2010 Jul 19
1
RMySQL and RMPI together
Dear R-users, I got problem when using both RMPI and RMySQL together and I thought some body in this forum can help me. I have a huge data set (more than 300,000 rows) and I would like to modify only few observations. I wrote the modifications in a function form and able to send the function to slaves using Robj2slaves from RMPI package. However the function needs MySQL data base name to extract
2007 Dec 20
2
Multicore computation in Windows network: How to set up Rmpi
R-users, My question is related to earlier posts about benefits of quadcore over dualcore computers; I am trying to setup a cluster of windows xp computers so that eventually I could make use of 10-20 cpu:s, but for learning how to do this, I am playing around with two laptops. I thought that the package snow would come handy in this situation, but to use snow, I would probably need to install
2010 Dec 08
0
copy old CF image to new CF device
I am using an 8G CF card. (I was using dd to duplicate the card with different size cards I was advised not to). So I wrote the script below which basically: 1) runs fdisk to setup the device 2) makes the ext3 file system and the swap 3) mounts my old image "/" filesystem and the new partition 4) tar and untar the directories 5) run grub on it 6) set label for "/" After I
2008 Nov 12
2
hotuser in multicore system
Incase of multi-core systems where multiple threads maybe scheduled for execution, does the below dscript (from hotuser in DTraceToolkit) collect information for multiple libraries/functions that would be getting executed at a given point of time. I have been able to test it to check if it does or doesn''t. I assume it does. This was a query I have about profiling in general. #pragma D
2011 Jun 06
1
parallel computing package on a multicore windows workstation
Hi, I would like to get suggestion about parallel computing package on a multicore windows workstation. I tried doSMP, but it crashes R a lot. I am wondering if "snow" and "snowfall" can be used on a single workstation (i,e, not cluster). At suggestion would be appreciated, Best, Richard [[alternative HTML version deleted]]
2010 Aug 12
1
multicore mclapply error
I'm running r 2. on a mac running 10.6.4 and a dual-core macbook pro. I'm having a funny time with multicore. When I run it with 2 cores, mclapply, R borks with the following error. The process has forked and you cannot use this CoreFoundation functionality safely. You MUST exec(). Break on __THE_PROCESS_HAS_FORKED_AND_YOU_CANNOT_USE_THIS_COREFOUNDATION_FUNCTIONALITY___YOU_MUST_EXEC__()
2011 Mar 29
0
Many cores support in R (Multicore Package)
Dear all, I am trying to improve my code for many cores. I have started with multicore package and the function mclapply. A multicore version of the lcapply. One problem I have is that when I use this function (you can copy and paste the below) require('multicore') returni <-function(i) {i} system.time(mclapply(seq(1:100000),returni)) I get 4 more versions of rkward.bin running (my
2012 Dec 28
1
Multicore/Parallel
I am using the package Multicore/Parallel to do importance sampling. I have 5 cores on my computer. And I have let's say 10 000 particles to generate. What I did was to send 5 particles in each time, calling the package parallel. Which means in all I am calling the parallel command 2000 times. What happens is in the end somewhere along the way I end up with the error message error in fork
2010 Jan 13
1
"select: bad file descriptor" in the multicore package
Hi, I wonder anyone knows what causes the error message "select: bad file descriptor" in the multicore package. This error sometimes occurs and sometimes doesn't. I couldn't find any documentation on this error about this package. thanks Jeff
2010 Jun 25
2
installing multicore package
Sir, I want to apply mclapply() function for my analysis. So, I have to install multicore package. But I can not install the package. >install.packages("multicore") It gives that package multicore is not available. Can you help me? Regards, Suman Dhara [[alternative HTML version deleted]]
2011 Jan 04
2
multicore support
hi to all, does wine support multicore cpus, yet? e.g. amd x2 or x4 processors thanks in advance
2010 Aug 04
0
[LLVMdev] regarding multicore support for LLVM
Multicore? You want OpenMP support? 2010/8/3 vijay kumar <vijaygbvv at gmail.com> > Hi all, > I am new to this LLVM. I went through the documenation of LLVM but > I didn't find any support for Multicore. Is there any such possibility where > multicore architecture can be exploited using LLVM. > > > > Thanks > Vijay > >
2010 Aug 04
0
[LLVMdev] regarding multicore support for LLVM
On Aug 3, 2010, at 8:48 PM, Liu wrote: > It is so difficult ... > Which FE? It need BE support? I didn't get it. > > 2010/8/4 vijay kumar <vijaygbvv at gmail.com> > Yeah OpenMP support. I read that it has a front end support but not the back end. So are there any projects or teams looking at this issue. > > On Wed, Aug 4, 2010 at 7:24 AM, Liu <proljc at