Displaying 20 results from an estimated 40 matches for "parallism".
2015 Sep 01
0
Switch to AIO instead of IO/parallize fs scan...
I was thinking about [Bug 3099]... in that while it's easy to get a 2-3x
speed
for the average app using parallel scans, the upper and lower bounds on that
speed increase could be <1x in a worst case (very unlikely, but with
primitive
or constrained (in a container or VM) HW, the chances are raised.
Better, with less std. deviation, I believe, might be to move I/O calls
to all being
AIO
2008 Dec 28
0
Roig Parals
This is a great wine from L'Empurd?, Spain. If you want to know more information contact:
Direction: C/ Garriguella 8
Town: Mollet de Peralada
Comarca: Alt Empord?
CP: 17752
Tel?fon: 972634320
It's the best wine I've ever seen.
2008 Nov 04
0
[LLVMdev] Available code-generation parallism
On Nov 3, 2008, at 3:55 PM, heisenbug wrote:
> What about "inventing" pseudo-constants (which point to the right
> thing) and build the piece of IR with them. When done, grab mutex and
> RAUW it in. Alternatively, submit to a privileged thread that performs
> the RAUW.
> The trick is to prepare the def/use chain(s) to a degree that the
> mutex is only held a minimal
2008 Nov 07
2
[LLVMdev] Available code-generation parallism
On Mon, 2008-11-03 at 01:06 -0800, Chris Lattner wrote:
> On Nov 2, 2008, at 2:20 PM, Jonathan Brandmeyer wrote:
> > I am interested in making my LLVM front-end multi-threaded in a way
> > similar to the GCC compiler server proposal and was wondering about
> > the
> > extent that the LLVM passes support it.
>
> Do you have a link for this? I'm not familiar
2006 Nov 20
5
Parallal Building?
I''m trying to index ~130,000 documents [soon to grow to about 500,000
documents] and I''m wondering if its possible to combine ferret databases
or in some other way split up the building process.
Normally, indexing 130k documents wouldn''t be that painful except that
there are different types of links between these documents and they are
not absolute (so for example
2008 Nov 03
2
[LLVMdev] Available code-generation parallism
On 3 Nov., 10:06, Chris Lattner <clatt... at apple.com> wrote:
> On Nov 2, 2008, at 2:20 PM, Jonathan Brandmeyer wrote:
>
> > I am interested in making my LLVM front-end multi-threaded in a way
> > similar to the GCC compiler server proposal and was wondering about
> > the
> > extent that the LLVM passes support it.
>
> Do you have a link for this?
2007 Oct 03
2
Running multiple instances of Samba and LDAP question
Dear All,
I work for the small IT company and I would like to ask the question concerning the project that I work on-Namely is it posible to run multiple instances of Samba and LDAP paralely on the one server --For example one cental server for Samba and LDAP and one instance running for each separate client -and if it possible the suggestion how to best accomplish that ...
I thank you in
2008 Nov 03
0
[LLVMdev] Available code-generation parallism
On Nov 2, 2008, at 2:20 PM, Jonathan Brandmeyer wrote:
> I am interested in making my LLVM front-end multi-threaded in a way
> similar to the GCC compiler server proposal and was wondering about
> the
> extent that the LLVM passes support it.
Do you have a link for this? I'm not familiar with any parallelism
proposed by that project. My understanding was that it was mostly
2008 Nov 02
4
[LLVMdev] Available code-generation parallism
I am interested in making my LLVM front-end multi-threaded in a way
similar to the GCC compiler server proposal and was wondering about the
extent that the LLVM passes support it.
Expression-at-a-time parallel construction:
If function definitions are built purely depth-first, such that the
parent pointers are not provided as they are created, what will break?
I noted that the function and module
2009 Jun 04
1
rsync --daemon. Can I open more than one instances?
Is there a way to speed up rsync by opening more than one daemon in
parallel. I use rsync --daemon to start rsync. I was wondering if
opening more than one instance is recommended or feasible? I know that
for services like nfs for example I have had opening many instances
improve performance.
I have 4 cores available so if there are any other parallization modes
I'd be glad to know since I
2003 May 05
2
(PR#1289)
Dear all,
I've found a bug report for R when installing it on an AIX5.1 system.
There's a followup (number 4) that comments a segmentation fault on R
when quitting "q()".
I'm using a newer version of R but on the same operating system and
I obtain the same error.
There was a way to solve this problem?
Best regards and thanks,
--
2012 Sep 16
2
Question about R performance on UNIX/LINUX with different memory/swap configurations
Does anyone have any guidance on swap and memory configuration when
running R v2.15.1 on UNIX/LINUX? Through some benchmarking across
multiple hardware (UNIX, LINUX, SPARC, x86, Windows, physical, virtual)
it "seems" that the smaller memory machines have an advantage.
Typically my organization builds their UNIX servers at a 1:1 physical
memory:swap configuration. We plan on running
2013 Jul 17
2
[LLVMdev] [Proposal] Parallelize post-IPO stage.
On 7/17/13 12:35 PM, Diego Novillo wrote:
> On Fri, Jul 12, 2013 at 3:49 PM, Shuxin Yang <shuxin.llvm at gmail.com> wrote:
>
>> 3. How to parallelize post-IPO stage
>> ====================================
>>
>> From 5k' high, the concept is very simple, just to
>> step 1).divide the merged IR into small pieces,
>> step 2).and compile
2013 Mar 11
0
[LLVMdev] AESOP autoparallelizing compiler
Hi Timothy,
Today I happened to download the code and do some experiments.
I actually wanted to see how you handle inter-procedure alias analysis.
So, I set inline threshold to zero and tried out following example
===============================================
#define N 1024
void func(double *A, double *B)
{
int i;
for (i=1; i<N-2; i++) {
B[i] = A[i] + i*3;
}
}
void func1(double
2013 Mar 11
1
[LLVMdev] AESOP autoparallelizing compiler
Hi Rahul,
Thanks for your interest!
Our work does not attempt to make any significant contributions to alias analysis, and acts as a client to existing LLVM AA. Furthermore, the options passed to the AESOP frontend scripts are obeyed at compile time, but at link time certain transformations occur unconditionally.
Here, AESOP has actually thwarted your experiment by performing inlining just
2013 Mar 10
2
[LLVMdev] AESOP autoparallelizing compiler
On Mon, Mar 04, 2013 at 03:01:15PM +0800, 陳韋任 (Wei-Ren Chen) wrote:
> Hi Timothy,
>
> > We would like to inform the community that we're releasing a version of our research compiler, "AESOP", developed at UMD using LLVM. AESOP is a distance-vector-based autoparallelizing compiler for shared-memory machines. The source code and some further information is available at
2013 Jul 17
0
[LLVMdev] [Proposal] Parallelize post-IPO stage.
On Wed, Jul 17, 2013 at 1:06 PM, Shuxin Yang <shuxin.llvm at gmail.com> wrote:
>
> On 7/17/13 12:35 PM, Diego Novillo wrote:
>>
>> On Fri, Jul 12, 2013 at 3:49 PM, Shuxin Yang <shuxin.llvm at gmail.com>
>> wrote:
>>
>>> 3. How to parallelize post-IPO stage
>>> ====================================
>>>
>>> From 5k'
2013 Mar 30
0
[LLVMdev] Problems with parallelizing lli
Hi,
I am trying to parallize the lli interpreter (code:
http://pastebin.com/6iuHNH3Q). I am using ubuntu 12.04 with llvm
version 3.1. Each thread uses a seperate LLVMContext, however the
interpreter continues to crash.
$ clang -S -emit-llvm test.c && parallel_lli test.s
Error msg: ..../include/llvm/Support/Mutex.h:116:
bool llvm::sys::SmartMutex<true>::release(): Assertion
2011 May 11
1
foreach(): how to do calculations "between" two foreach loops?
Dear expeRts,
is it possible to carry out calculations between different foreach() calls?
As for nested loops, you want to carry out calcuations not depending on the inner
loop only once and not for each iteration of the innermost loop.
Cheers,
Marius
library(foreach)
foreach(i=1:3) %:%
foreach(j=1:2) %do% {
i <- i+1
print(paste(i,j))
}
foreach(i=1:3) %:%
i
2011 Jul 22
0
Mclapply prints once
Dear all
I have the following code that works in paraller (and is pretty fast actually)
The only problem I still have is that every core just prints only once.
dimz<-1000
Shadowlist<-mclapply(1:dimz, function(i) {
????????????????????????????? print(sprintf('Creating the %d map',i));
????????????????????????????? createField(x=x, y=y, model=model,