similar to: Rmpi 0.5-6 : error spawning process

Displaying 20 results from an estimated 800 matches similar to: "Rmpi 0.5-6 : error spawning process"

2007 Sep 03
1
Snow on Windows Cluster
Hello, the package snow is not working on a windows cluster with MPICH2 and Rmpi. There is an error in makeCluster: launch failed: CreateProcess(/usr/bin/env "RPROG="C:\Programme\R\R-2.5.1\bin\R" "OUT=/dev/null" "R_LIBS=" C:/Programme/R/R-2.5.1/library/snow/RMPInode.sh) on 'cl1' failed, error 3 - Das System kann den angegbenen Pfad nicht finden. I
2012 Apr 04
1
npRmpi trouble - mpi.comm.spawn causes segfault
Dear all, I have a large dataset of randomly generated weighed sample for which I wish to compute a kernel density estimate. I have used the "np" package successfully for smaller datasets, however for the larger ones, they take too long when using the cross validation options for bandwidth selection ("cv.ls" or "cv.ml"). Of course, they are much quicker with
2008 Sep 27
1
Problem with R on dual core under Linux - can not execute mpi.spawn.Rslaves()
Hi I am trying to utilize my dual core processor (and later a High-performance clusters (HPC) ) by using the Rmpi, snow, snowfall, ... packages, but I am struggling at the beginning, i.e. to initialise the "cluster" on my dual core computer. Whenever I try to initialize it (via sfInit(parallel=TRUE, cpus=2) or mpi.spawn.Rslaves(nslaves=2) ), I get an error message: >
2008 Mar 20
1
Rmpi and C Code, where to get the communicator
Hello, I try to write parts of my code in C to accelerate the for-loops. But basic operations I want to do in R (e.g. start cluster). My R code looks something like this: library(Rmpi) mpi.spawn.Rslaves() mpi.remote.exec(....) dyn.load("test.so") erg <- .Call("test", ....) .... mpi.close.Rslaves() mpi.quit() And my C function looks something like this: #include
2008 Mar 27
1
snow, stopping cluster
Hello, is there any function in the package snow to check for a really running cluster? The function checkCluster only checks the variable cl. And the variable is still available after stopping the cluster! ( a simple solution would be deleting the cluster variable cl in the function stopCluster) > library(snow) > cl <- makeCluster(5) 5 slaves are spawned successfully. 0
2008 Oct 22
1
Problem about spawn nodes with Rmpi
Hi all, now I'm testing R in a "virtual cluster", made it with VirtualBox. This one has 3 nodes, running CentOS 5 and OpenMPI 1.2.8, and the principal node (called "server") exports the /home to other nodes. I have installed R and OpenMPI in /home, in fact, it seems work OK. Editing the openmpi-default-hostfile and run "mpirun -np 3 hostname" I can see the
2010 Jul 12
1
How to use mpi.allreduce() in Rmpi?
Hi everybody! I have the next code which makes a reduction of the *a *variable in two slaves, using the Rmpi package. library(Rmpi) mpi.spawn.Rslaves(nslaves=2) reduc<-function(){ a<-mpi.comm.rank()+2 mpi.reduce(a,type=2, op="prod") return(paste("a=",a)) } mpi.bcast.Robj2slave(reduc) mpi.remote.exec(reduc()) cat("Product: ")
2010 Jul 12
1
How to use mpi.allreduce() in Rmpi?
Hi everybody! I have the next code which makes a reduction of the *a *variable in two slaves, using the Rmpi package. library(Rmpi) mpi.spawn.Rslaves(nslaves=2) reduc<-function(){ a<-mpi.comm.rank()+2 mpi.reduce(a,type=2, op="prod") return(paste("a=",a)) } mpi.bcast.Robj2slave(reduc) mpi.remote.exec(reduc()) cat("Product: ")
2008 Nov 07
1
Rmpi task-pull
Hi, I'm testing the efficiency of the Rmpi package regarding parallelization using a cluster. I've found and tried the task pull programming method, but even if it is described as the best method, it seems to cause deadlock, anyone could help me in using this method? here is the code I've found and tried: # Initialize MPI library("Rmpi") # Notice we just say "give us
2008 Jul 10
0
Rmpi unkown input format error
I have just installed Rmpi on a Suse 9.1 linux cluster with openmpi-1.0.1. I am trying the example included below from the tutorial website. However, I keep getting the following error: > # Load the R MPI package if it is not already loaded. > if (!is.loaded("mpi_initialize")) { + library("Rmpi") + } > > # Spawn as many slaves as possible >
2008 Oct 15
0
R-help Digest, Vol 67, Issue 31
V; Sent via BlackBerry from T-Mobile -----Original Message----- From: r-help-request at r-project.org Date: Tue, 30 Sep 2008 12:00:06 To: <r-help at r-project.org> Subject: R-help Digest, Vol 67, Issue 31 Send R-help mailing list submissions to r-help at r-project.org To subscribe or unsubscribe via the World Wide Web, visit https://stat.ethz.ch/mailman/listinfo/r-help or, via email,
2007 Jul 05
0
Question on Rmpi looping
Dear R list, In the course of learning to work with Rmpi, we are confused about a few points. The following simple program is based on some examples we retrieved from the web. Each slave is writing the same output line multiple times (a multiple equal to the number of slaves). In other words, the write statements are being executed a number of times equal to the number of slaves. I am
2011 Feb 01
1
Rmpi; sample code not running, the slaves won't execute commands
Hi All, I'm trying to parallelize some code using Rmpi and I've started with a sample 'hello world' program that's available at http://math.acadiau.ca/ACMMaC/Rmpi/sample.html. The code is as follows; # Load the R MPI package if it is not already loaded. if (!is.loaded("mpi_initialize")) { library("Rmpi") } # Spawn as many slaves as possible
2007 Sep 07
2
ploting missing data
Hello, I have this kind of dataframe and have to plot it. data <- data.frame(sw= c(1,2,3,4,5,6,7,8,9,10,11,12,15), zehn = c(33.44,20.67,18.20,18.19,17.89,19.65,20.05,19.87,20.55,22.53,NA,NA,NA), zwanzig = c(61.42,NA,26.60,23.28,NA,24.90,24.47,24.53,26.41,28.26,NA,29.80,35.49), fuenfzig =
2007 Aug 06
1
Problems with expresso
Hello, I want to use expresso for preprocessing the hgu133a-spikein data from affycompII. But there is an error: > library(affy) > path <- "z:/Microarray/hgu133a-spikein/rawdata" > celFile <- list.celfiles(path=path,full.names=TRUE); > affyBatch <- ReadAffy(filenames=celFile[1:6]); > eset1 <-
2007 Jan 19
5
Problem with C extension
Hello, I try to write an extension in C, to get a faster functions. Therefore I have to add an element (vector) to a vector. The command in R is very simple: x = c(x,a) But in C I have the problem to reallocate my vector for getting more space. Everything I tried, I get a "Segmentation fault". So, how can I combine two vectors in C and give the result back to R (return(x))? Thanks
2010 Oct 04
0
Syntax for Rmpi cf multicore
I'm aiming to compare the workings of Rmpi and multicore on a duel processor quad core machine with 64 bit R-2.11.1 Kubuntu 10.4. It's impossible for me to get a small reproducable code segment to show what I mean, but if I show what works for mclapply, I'd hope it's possible to be shown what would be the equivalent way with mpi.apply. The function lr.gbm has variables trees,
2007 Aug 14
0
papply, reuse of data
Hello, we are doing some computations with MPI and Rmpi. Therefore papply is a very comfortable function. Is it possible to reuse calculated data on slaves? We do some calculation with papply, then we have to do some overall calculation at the master then we want to do some calculation on the slaves with the old results and some new results from the master. We do not want to send all data
2007 Jun 11
3
simultaneous computing
Hello, which possibilities are available in R for simultaneous or parallel computing? I only could find biopara (http://cran.r-project.org/src/contrib/Descriptions/biopara.html) Are there other possibilities? Are there special groups working on simultaneous computing with R? Thanks Markus -- Dipl.-Tech. Math. Markus Schmidberger Ludwig-Maximilians-Universit?t M?nchen IBE - Institut f?r
2007 Mar 28
4
Rmpi and OpenMPI ?
Has anybody tried to use Rmpi with the OpenMPI library instead of LAM/MPI? LAM appears to be somewhat hardcoded in the Rmpi setup. Before I start to experiment with changing this, has anybody else tried Rmpi with non-LAM MPI implementations? Dirk -- Hell, there are no rules here - we're trying to accomplish something. -- Thomas A. Edison