Displaying 20 results from an estimated 400 matches similar to: "[LLVMdev] -gcolumn-info and PR 14106"
2014 Jun 26
2
[LLVMdev] -gcolumn-info and PR 14106
On Thu, Jun 26, 2014 at 12:39 PM, Robinson, Paul
<Paul_Robinson at playstation.sony.com> wrote:
> The main motivation for turning it off is that no known consumer (debugger)
> took advantage of it.
>
There are lots of non-debugger consumers of debug info, and now we've
got cases where we actively want it for those.
> Turning it on does more than slightly increase the object
2014 Jun 26
2
[LLVMdev] -gcolumn-info and PR 14106
On Thu, Jun 26, 2014 at 3:39 PM, Robinson, Paul
<Paul_Robinson at playstation.sony.com> wrote:
> The main motivation for turning it off is that no known consumer (debugger)
> took advantage of it.
>
> Turning it on does more than slightly increase the object file size, it can
> cause the same source line to be listed multiple times in the .debug_line
> table (with different
2018 Nov 28
2
Finding line numbers in source code associated with given bitcode instructions
Hello,
Is it possible to somehow similar to how DWARF functions with assembly find
out which line numbers in unoptimized code are associated with which
bitcode instructions for the purpose of debugging compiler passes? I am
curious about whether this is possible prior to writing some
instrumentation passes for the purpose of debugging them.
Thanks in advance,
Carter.
-------------- next part
2014 Aug 28
2
[LLVMdev] Minimizing -gmlt
On Thu, Aug 28, 2014 at 11:51 AM, Alexey Samsonov <vonosmas at gmail.com>
wrote:
> This sounds great. Teaching backend about the -gmlt might help us in
> another way: we might enforce full debug info generation in the frontend
> for -fsanitize= flags, then rely on some parts of this debug info in
> instrumentation passes and prune it before the actual object file
>
2014 Aug 27
6
[LLVMdev] Minimizing -gmlt
In an effort to fix inlined information for backtraces under DWARF Fission
in the absence of the split DWARF (.dwo) files, I'm planning on adding
-gmlt-like data to the .o file, alongside the skeleton CU.
Since that will involve teaching the LLVM about -gmlt (moreso than it
already has - the debug info LLVM metadata already describes -gmlt for the
purposes of omitting pubnames in that case) I
2018 Aug 14
2
optimization remarks
Hi,
I am trying to compare the loop vectorizers effectiveness for different
targets relative to each other. That way, I am hoping to find loops that
are not vectorized - but could be - on my target by finding other
targets doing this successfully. With some luck, there might be
something in the Target files that could be fixed with improved
vectorization as a result...
I would like to do
2019 May 02
2
llvm is illegally vectorizing with a recurrence on skylake
Hi -- I have found a bug in an HPC code where llvm is vectorizing a loop on
Skylake that has an obvious recurrence. I derived a small test case based
on the original benchmark below:
/*****************************************************************/
static void __attribute__ ((always_inline)) one(
const int *restrict in, const int *const end,
const unsigned shift, int *const restrict index,
2018 Jun 05
2
How to get optimization remarks while testing with lnt in llvm
Hi, I'm new to llvm and am trying to run benchmarks from the test-suite
using lnt to check loop-vectorization for various benchmarks.
Test are compiling and executing fine, but I am not getting optimization
remarks while using flags like -Rpass-missed=loop-vectorize and
-Rpass-analysis=loop-vectorize
I've tried running it like this:
lnt runtest test-suite --sandbox SANDBOX --cc
2016 Mar 07
3
Profile-based inlining status
Hello,
I'm learning how LLVM performs PGO (profile-guided optimizations) by using
the instrumentation-based profile build (-fprofile-instr-generate and
-fprofile-instr-use).
However, I found there is no difference in inlining behaviors between with
and without PGO for a few spec benchmarks by checking the emit optimization
reports (-Rpass=inline -Rpass-missed=inline -Rpass-analysis=inline).
2018 Feb 12
1
Pattern not recognized as reduction
Reduction Not Captured By LLVM
CODE_1
------------------------------------------------------------
------------------------------------------------------------
--------------------
#include <stdio.h>
int main()
{
int sum[1000]={1,2,3,4};
for (int i=1;i<1000;i++)
{
sum[0] +=sum[i-1];
}
}
2016 Oct 09
3
On Loop Distribution pass
Dear community,
Our team at IITH have been experimenting with loop-distribution pass in
LLVM. We see the following results on few benchmarks.
clang -O3 -mllvm -enable-loop-distribute -Rpass=loop-distribute file.c
clang -O3 -mllvm -enable-loop-distribute -Rpass-analysis=loop-distribute
file.c
TORCH
2016 May 11
4
Filter optimization remarks by the hotness of the code region
> On May 11, 2016, at 3:37 AM, Hal Finkel <hfinkel at anl.gov> wrote:
>
> ----- Original Message -----
>> From: "Adam Nemet" <anemet at apple.com>
>> To: "Hal Finkel" <hfinkel at anl.gov>
>> Cc: "llvm-dev (llvm-dev at lists.llvm.org)" <llvm-dev at lists.llvm.org>
>> Sent: Wednesday, May 11, 2016 1:15:42 AM
2018 Sep 20
2
Vectorization width not correct using #pragma clang loop vectorize_width
Hello,
I m trying to set vector width using #pragma clang loop vectorize_width(32)
but i m getting width 8 for the following kernel;
#define M 128
#define N 128
#define SQRT_FUN(x) sqrtf(x)
int main(int argc, char** argv)
{
/* Variable declaration/allocation. */
double float_n = (double)N;
double data[N*M];
double corr[M*M];
double mean[M];
double stddev[M];
uint32_t
2020 Jun 24
2
Loop vectorization and unsafe floating point math
Hi llvm-dev!
We are doing some fuzzy testing using C program generators,
and one question that came up when generating a program with
both floating point arithmetic and loop pragmas was;
Is the loop vectorizer really allowed to vectorize a loop when
it can't prove that it is safe to reorder fp math, even if
there is a loop pragma that hints about a preferred width.
When reading here
2016 May 04
4
Filter optimization remarks by the hotness of the code region
This idea came up a few times recently [1][2] so I’d like start prototyping it. To summarize, we can emit optimization remarks using the -Rpass* options. These are currently emitted by optimizations like vectorization[3], unrolling, inlining and since last week loop distribution.
For large programs however this can amount to a lot of diagnostics output to sift through. Filtering this by the
2020 Sep 01
2
Vectorization of math function failed?
I've tried to do:
clang++ -O3 -march=native -mtune=native \
-Rpass=loop-vectorize,slp-vectorize
-Rpass-missed=loop-vectorize,slp-vectorize
-Rpass-analysis=loop-vectorize,slp-vectorize \
-ffast-math -ffp-model=fast -ffp-exception-behavior=ignore -ffp-contract=fast \
-c -o vec.o vec.cc
But I've got no feedback.
--
Alexandre Bique
2015 Feb 09
3
[LLVMdev] aarch64 status for generating SIMD instructions
% clang -S -O3 -mcpu=cortex-a57 -ffast-math -Rpass-analysis=loop-vectorize dot.c
dot.c:15:1: remark: loop not vectorized: value that could not be identified as
reduction is used outside the loop [-Rpass-analysis=loop-vectorize]
}
^
dot.c:15:1: note: could not determine the original source location for :0:0
I found “llvm-as < /dev/null | llc -march=aarch64 -mattr=help” which listed a
2017 Oct 18
2
How to emit opt report when using LTO
Hi,
I'm using clang frontend.
I'm interested in some particular hot loop in my code and I emit a
report from vectorizer optimizations passes.
I receive nice output if passing -Rpass* flags as long as I'm building
without LTO?
But with -flto it just prints nothing.
Is there a way to emit opt reports when using LTO?
For now I can only approximate about whether my the loop will be
2016 May 11
2
Filter optimization remarks by the hotness of the code region
Hi Hal,
> On May 10, 2016, at 5:39 PM, Hal Finkel <hfinkel at anl.gov> wrote:
>
> Hi Adam,
>
> I think would be a really useful feature to have. I don't think that the backend should be responsible for filtering, but should pass the relative hotness information to the frontend. Given that these diagnostics are not just going to be used for -Rpass and friends, but also
2016 Mar 02
4
Proposal for function vectorization and loop vectorization with function calls
Proposal for function vectorization and loop vectorization with function calls
==============================================================================
Intel Corporation (3/2/2016)
This is a proposal for an initial work towards Clang and LLVM implementation of
vectorizing a function annotated with OpenMP 4.5's "#pragma omp declare simd"
(named SIMD-enabled function) and its