search for: parallelizable

Displaying 20 results from an estimated 81 matches for "parallelizable".

2011 Apr 20
3
[LLVMdev] Coarse-grained parallelism
Am 19.04.2011 um 16:44 schrieb John Criswell: > On 4/19/11 5:57 AM, Andreas Wilhelm wrote: >> >> Hello, >> >> I found some code within the pool allocation project to identify parallelizable function calls. >> Unfortunately the functionality isn't part of the current release of poolalloc (in release 14 it was). > > Can you tell me in what file(s) this is implemented? I wasn't aware that the poolalloc project had such an analysis. The automatic parallelization was...
2011 Apr 19
2
[LLVMdev] Coarse-grained parallelism
Hello, I found some code within the pool allocation project to identify parallelizable function calls. Unfortunately the functionality isn't part of the current release of poolalloc (in release 14 it was). My intention is to estimate the parallelization-potential of sequential applications concerning coarse-grained parallelism. Can you tell me... 1. Why are classes of polla...
2011 Apr 19
0
[LLVMdev] Coarse-grained parallelism
On 4/19/11 5:57 AM, Andreas Wilhelm wrote: > Hello, > > I found some code within the pool allocation project to identify > parallelizable function calls. > Unfortunately the functionality isn't part of the current release of > poolalloc (in release 14 it was). Can you tell me in what file(s) this is implemented? I wasn't aware that the poolalloc project had such an analysis. > > My intention is to estimate th...
2011 Apr 22
0
[LLVMdev] Coarse-grained parallelism
On 04/20/2011 08:05 AM, Andreas Wilhelm wrote: > Am 19.04.2011 um 16:44 schrieb John Criswell: > >> On 4/19/11 5:57 AM, Andreas Wilhelm wrote: >>> Hello, >>> >>> I found some code within the pool allocation project to identify >>> parallelizable function calls. >>> Unfortunately the functionality isn't part of the current release of >>> poolalloc (in release 14 it was). >> >> Can you tell me in what file(s) this is implemented? I wasn't aware >> that the poolalloc project had such an analysis. &g...
2016 Nov 16
9
LLD: time to enable --threads by default
...l need to focus on single-thread performance rather than multi-threaded one because it is hard to make a slow program faster just by using more threads. - We shouldn't do "too clever" things with threads. Currently, we are using multi-threads only at two places where they are highly parallelizable by nature (namely, copying and applying relocations for each input section, and computing build-id hash). We are using parallel_for_each, and that is very simple and easy to understand. I believe this was a right design choice, and I don't think we want to have something like workqueues/tasks i...
2017 Jan 03
2
Automatic Insertion of OpenACC/OpenMP directives
...[5]],y[0:AI1[5]]) >> #pragma acc kernels >> for (int i = 0; i < n; ++i) { >> y[j] = a * x[i] + y[j]; >> ++j; >> } > > I'm not familiar with OpenACC, but doesn't this still have a loop carried dependence on j, and therefore isn't correctly parallelizable as written? That was my original concern as well, but I had forgot that OpenACC pragma are not necessarily saying to the compiler that the loop is parallel: #pragma acc kernels only tells the compiler to “try” to parallelize the loop if it can prove it safe, but: #pragma acc parallel kernels...
2016 Nov 23
3
LLD: time to enable --threads by default
...rather than >> multi-threaded one because it is hard to make a slow program faster just by >> using more threads. >> >> - We shouldn't do "too clever" things with threads. Currently, we are >> using multi-threads only at two places where they are highly parallelizable >> by nature (namely, copying and applying relocations for each input section, >> and computing build-id hash). We are using parallel_for_each, and that is >> very simple and easy to understand. I believe this was a right design >> choice, and I don't think we want to ha...
2011 Apr 27
1
[LLVMdev] Coarse-grained parallelism
...2011 08:05 AM, Andreas Wilhelm wrote: >> Am 19.04.2011 um 16:44 schrieb John Criswell: >> >>> On 4/19/11 5:57 AM, Andreas Wilhelm wrote: >>>> Hello, >>>> >>>> I found some code within the pool allocation project to identify >>>> parallelizable function calls. >>>> Unfortunately the functionality isn't part of the current release of >>>> poolalloc (in release 14 it was). >>> >>> Can you tell me in what file(s) this is implemented? I wasn't aware >>> that the poolalloc project had...
2019 Feb 15
4
Can we disable diffie-hellman-group-exchange-sha1 by default?
I'm not nearly knowledgeable enough in crypto to fully understand your answer, but I will try. I wonder why moduli are not automatically generated the first time sshd is started though. That would make much more sense than shipping a default moduli file but also asking everyone to replace it with their own. On Fri, Feb 15, 2019 at 5:50 AM Mark D. Baushke <mdb at juniper.net> wrote: >
2004 Jan 08
1
[LLVMdev] Re: idea 10
Hello Valery, I have some comments regarding your thoughts on LLVM support for distributed computing. Valery A.Khamenya wrote: >There should be an engine and layer for making dispatching optimizations in run-time. If one CPU is loaded and code is >"parallelizable" why then not to send some part of >calculation to other CPU? This kind of on-fly decision will >be one day incorporated in something like LLVM. > > I'm not sure to correctly understand what you mean, but I interpret it as LLVM deciding where the code should be executed, l...
2019 Feb 15
2
Can we disable diffie-hellman-group-exchange-sha1 by default?
On Fri, 2019-02-15 at 15:57 +1100, Darren Tucker wrote: > That was the original intent (and it's mentioned in RFC4419) however > each moduli file we ship (70-80 instances of 6 sizes) takes about 1 > cpu-month to generate on a lowish-power x86-64 machine. Most of it > is > parallelizable, but even then it'd likely take a few hours to > generate > one of each size. I imagine that'd cause some complaints about > startup time. One way of handling this, at least if the moduli generation was nicely interruptable, is that distros ship a file, e.g. /etc/ssh/moduli.dist...
2012 Aug 08
1
Question about R and multiple CPU's
I have a question about multiple cores and CPU's for running R. I've been running various tests on different types of hardware and operating systems (64 bit, 32 bit, Solaris, Linux, Windows, RV.10, .12, .15, .15.1.) Generally speaking, it seems that for a single user and process that R prefers to have as much resources as possible; especially memory. I've looked at some of the r-sig
2008 Mar 20
0
[LLVMdev] Array Dependence Analysis
...for corner cases. If there is an interest I can share/contribute it soon. However, I wouldn't expect much from it and would rather treat it as a toy. After your advice, I am going to prepare (locally) custom tests for llvm-test testsuite. The number of loops not carrying memory dependences, or parallelizable ones can be a good measure of dependence analysis precision. Comparing output of parallelized programs to the original ones would, probably, help to check its correctness. > If you're looking for a simple client, I'd suggest something like loop > reversal. I think it would exercis...
2019 Feb 15
3
Can we disable diffie-hellman-group-exchange-sha1 by default?
...also asking > > everyone to replace it with their own. > > That was the original intent (and it's mentioned in RFC4419) however > each moduli file we ship (70-80 instances of 6 sizes) takes about 1 > cpu-month to generate on a lowish-power x86-64 machine. Most of it is > parallelizable, but even then it'd likely take a few hours to generate > one of each size. I imagine that'd cause some complaints about > startup time. > > With those caveats, you are also welcome to add the appropriate > ssh-keygen commands to your startup scripts. > > -- > Darre...
2017 Jun 05
1
more recent perl version?
...be pulled in by the page proper.) > Perl is pretty fast, and most of the work is being done by the database, > so I?m not sure how using an alternative to CGI could make things go faster. There are many reasons CGI is relatively slow. 1. If you have many connections per second, you need a parallelizable technology to make things go fast. If everything chokes down to a single thread, you?ve got a serious bottleneck when the CPS count gets high enough. 2. CGI re-launches the cgi-bin/* programs on every page hit. Since Perl doesn?t leave behind ?compiled? forms of the program as some other dynamic...
2008 Nov 30
2
Snow and multi-processing
Dear R gurus, I have a very embarrassingly parallelizable job that I am trying to speed up with snow on our local cluster. Basically, I am doing ~50,000 t.test for a series of micro-array experiments, one gene at a time. Thus, I can easily spread the load across multiple processors and nodes. So, I have a master list object that tells me what rows to pic...
2004 Jan 08
0
[LLVMdev] Re: idea 10
...o the set of languages low level primitives and constructs. if you don't pollute Fib. function with explicit optimizations and use language like C, then what kind of level should bring "parallel optimization" to your Fib. code?.. don't forget, I'd like to parallelize any parallelizable code, like in this Fib. example, written in C > I personally don't think that automating the conversion of an algorithm > from non-distributed to distributed execution can be done at a low > level, mainly because it involves many high-level constructs. well, don't kill me, bu...
2012 Feb 04
1
[LLVMdev] [BBVectorizer] Obvious vectorization benefit, but req-chain is too short
...er r149761). If this "solution" causes other problems, > then we may need to think of something more sophisticated. I wonder if the case where a store is the last user of the value could be treated as a special case. The case where the code reads, computes and writes values in a fully parallelizable (unrolled) loop is an optimal case for vectorizing as it might not lead to any unpack/pack overheads at all. In case of the bb-vectorizer (if I understood the parameters correctly), if the final store (or actually, any final consumer of a value) is weighed more heavily in the "chain length co...
2016 Dec 31
2
Automatic Insertion of OpenACC/OpenMP directives
Dear Mehdi, I've changed your example a little bit: float saxpy(float a, float *x, float *y, int n) { int j = 0; for (int i = 0; i < n; ++i) { y[j] = a*x[i] + y[I]; // Change 'I' into 'j'? ++j; } } I get this code below, once I replace 'I' with 'j'. We are copying n positions of both arrays, 'x' and 'y': float saxpy(float a,
2008 Mar 19
2
[LLVMdev] Array Dependence Analysis
On Mar 18, 2008, at 9:21 AM, Wojciech Matyjewicz wrote: > Hi, > >> Cool! I think the most critical part of this is to get a good >> interface for dependence analysis. There are lots of interesting >> implementations that have various time/space tradeoffs. >> >> For example, it would be great if Omega was available as an option, >> even if the compiler