Displaying 20 results from an estimated 10000 matches similar to: "clang dwarf"
2018 May 22
2
LLVM dwarf info is not complete
Hi
I am using llvm-dwarfdump to dump the line table with -debug-line option. I
compiled my program with -gdwarf-3.
I first get the line number and source number for every instruction of the
LLVM IR with metadata. Then I try to find the corresponding binary address
from the dwarf debug info. However, I noticed that the dwarf table might
not always return me the answer and some of the source line
2018 May 22
0
LLVM dwarf info is not complete
> On May 22, 2018, at 8:06 AM, Muhui Jiang via llvm-dev <llvm-dev at lists.llvm.org> wrote:
>
> Hi
>
> I am using llvm-dwarfdump to dump the line table with -debug-line option. I compiled my program with -gdwarf-3.
>
> I first get the line number and source number for every instruction of the LLVM IR with metadata. Then I try to find the corresponding binary address
2018 Jun 26
2
Instruction boundaries
There should be a line-table entry for the end of the function, which appears to be missing from the dump you provided. llvm-dwarfdump should report this address with 'end_sequence' in the Flags. Are you using a different dumper?
I am not sure but my guess would be that inline data is not represented in the line table. The line table's primary purpose is to inform the debugger
2018 Jun 26
2
Instruction boundaries
I'm not familiar with the target instruction set, but if "MOV PC, R0" is not a return instruction, I'm guessing that the sequence starting at A39C is a dispatch through a jump table. The jump table would be considered part of the instruction stream and included in the scope of the line table. This is not a case where you would see end_sequence; my mistake.
The line table does
2018 Jun 26
2
Instruction boundaries
Hi paulr
Thanks for your reply. Though DWARF info give me the code address ranges,
there might be inline data. If so, how to handle this case?
As for the dwarf line table. Sometimes, the source line might be zero. Do
you know why? If all instructions should be describe in the line table, I
think analyzing Dwarf line table is enough to get all the instructions
addresses. Do you agree?
I would
2018 Jun 13
2
IR to binary address mapping
Hi Paul
Thanks for your comments. Suppose I can generate the control flow graph via
LLVM Pass or the default option like '-dot-cfg' with opt. However, the
control flow graph is based on llvm IR level. I would like to have a
control flow graph based on binary level. Thus, I want to map the IR to
binary address.
As far as I know, we used to use the debug information to map the IR to
source
2018 Jun 03
2
Function start address
Hi Muhui,
I tried to grep the "DW_TAG_subprogram" from the debug_info . However, I noticed that the number I found is still less than the whole functions I found with LLVM IR. Do you have any experiences? Many Thanks
The only explanation that comes to mind, is that the functions are not in the final binary object file. However, previously you said you believed they were present. If
2018 Jun 13
2
IR to binary address mapping
Hi
However, frontend may also do various operations on the source code and one
line number and column number could map to more than one binary address.
Why LLVM IR cannot?
Regrads
Muhui
2018-06-12 23:18 GMT+08:00 mayuyu.io <admin at mayuyu.io>:
> In theory that’s not exactly possible/accurate. Due to various operations
> in the Backend like Instruction Legalization, one IR
2018 Jun 02
2
Function start address
Hi
Actually, No particular reason. I just think this might be a solution, then
I use think kind of method. Querying the symbol table would be a good
choice, but I prefer to use LLVM and dwarf information. I am sorry that I
am not familiar with debug_info. But thanks to your suggestions. I would
like to try to solve it with debug_info. It seems work according to your
comments
By the way, I am
2018 Jun 25
2
Instruction boundaries
Hi
I was wondering whether there are any methods to know what part of the
target binary is code.
I have some ideas and hope to get your comments.
I would like to use LLVM's source level debugging information to extract
the source lines belonging to every functions. Then use the dwarf mapping
table to transfer the source level information to binary address. Are
there any better methods?
2018 Jun 12
4
IR to binary address mapping
Hi
I know that LLVM provide some debug API for us to know the source code
information. For example, every IR instruction's source line number and
column number.
However, are there any method to get a mapping from IR instruction to
binary address directly. I don't want to use dwarf line mapping table as a
bridge. I think the binary is generated by clang and llvm. I think there
definitely
2018 Sep 29
2
Function Signature
Hi
I was wondering whether I can extract the function signature from the dwarf
debugging information or symbol table. I mean the argument list and the
return variables.
I know that in dwarf debugging information, a subprogram represents a
function. However, I could not find the information that is related to the
function signature. Any ideas? Many Thanks
Regards
Muhui
-------------- next part
2018 Jun 01
3
Function start address
Hi
I am using LLVM Pass combined with dwarf debug information to get all the
function's start address. My steps are below:
First, I write the function pass to get the start line of each function,
which is finished.
Then, based on the start line of every single function, I try to query the
specific line from the dwarf's line binary table, which is generated with
llvm-dwarfdump
2018 Nov 03
2
llvm bug 36466 fix
Hi Dave
Sorry, I meant the hardware you're using to compile LLVM - you mentioned it
took you a long time to rebuild it so it would be hard for you to
write/experiment on tests.
=============================
Compiling LLVM doesn't take me too much time(less than 2 hours). The
hardware is good enough and I am using interl E5 CPU. What I mean is that
it took me a long time to analysis the
2018 May 15
1
Four bitcode generated with plugin-opt=save-temps
Hi Teresa
Thanks for your very quick and clear explanation. I have one more question.
The emit-llvm option will give you the IR for a single source file when you
compile it with -c. All of those files when combined give the IR in the
preopt.bc temp file.
===========
So if I use "clang -emit-llvm -c" to generate the .ll file. It should be
the same as the one I generated by using
2018 Nov 03
2
llvm bug 36466 fix
Hi Dave
I am not going to access any hardware. I am using clang to analysis the ARM
binaries. The binary is 483.xalancbmk in CPU SPEC2006. When I use the
optimization O0, no crash will occur. The crash occurs when I set
optimization level as O1,O2,O3 and Os.
If I have to recompile and rerun the tests. What version of llvm is
suggested. It would be better if anyone could provide the patch on this
2018 May 16
0
clang llvm cross compile with autotools
Hi
Now I am using clang and llvm to do cross compiling. My target architecture
is arm and the host is x86-64. I am trying to compile coreutils and
SPEC2006(Anyone who ever has the same experience, please tell me. Really
need your help)
At the same time, I hope to generate the LLVM IR, which I can write PASS to
extract some information. I would also need the dwarf debug information.
I briefly
2018 May 15
2
Four bitcode generated with plugin-opt=save-temps
Hi
I use the LDFLAGS=" -flto -fuse-ld=gold -Wl,-plugin-opt=save-temps " to
generate the makefile and to make the whole program. However, found four
different kinds of bitcode for each target. For example, I am compiling
coreutils. For the program "nohup", I can get
nohup.0.0.preopt.bc
nohup.0.2.internalize.bc
nohup.0.4.opt.bc
nohup.0.5.precodegen.bc
If I am right, I
2018 May 15
0
Four bitcode generated with plugin-opt=save-temps
These are the bitcode at different stages of the LTO portion of the
compile. LTO merges the IR for all files being linked and optimizes them as
a single monolithic module. The preopt.bc is the merged IR just after
merging and before performing any LTO optimizations. internalize.bc is
after performing whole program internalization. opt.bc is after the
optimization pipeline, and .precodegen.bc is
2018 Jul 01
2
Cross Compilation Problem
Hi Peter
I guess this is the document written by you
https://fosdem.org/2018/schedule/event/crosscompile/attachments/slides/2107/export/events/attachments/crosscompile/slides/2107/How_to_cross_compile_with_LLVM_based_tools.pdf
I follow it to try to use the clang to do cross compilation. Actually, my
target binaries is SPECCPU2006 and autotools based binaries.
However, I failed on the first