Ben Shi via llvm-dev
2020-Jun-30 03:36 UTC
[llvm-dev] How to prevent llvm's default optimization
Hi, James, Thanks for your reply. I do not think it is always true, that "mul then add" is faster than "add then mul". For example, A small immediate can be directly encoded in the instruction, but it becomes a larger one after a multiplication, which has to be loaded from the constant pool (extra memory access). So I wonder, is it possile to prevent it, via changes to the config of the base class TargetLowering,than writing special custom C++ code. Ben At 2020-06-30 07:46:41, "James Courtier-Dutton" <james.dutton at gmail.com> wrote:>Hi Ben, > >Why do you want to stop it? >"mul then add" is faster than "add then mul". >The result is the same in both cases. > >On Mon, 29 Jun 2020 at 22:11, Ben Shi via llvm-dev ><llvm-dev at lists.llvm.org> wrote: >> >> Hello, >> I have an instruction pattern like >> %2 = add i32 %0, 25 >> %3 = mul i32 %2, 525 >> >> and llvm will optimize it to >> %2 = mul i32 %0, 525 >> %3 = add i32 %2, 12525 >> >> how to prevent it? >> >> _______________________________________________ >> LLVM Developers mailing list >> llvm-dev at lists.llvm.org >> https://lists.llvm.org/cgi-bin/mailman/listinfo/llvm-dev-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.llvm.org/pipermail/llvm-dev/attachments/20200630/cb1b12cd/attachment.html>
Craig Topper via llvm-dev
2020-Jun-30 03:53 UTC
[llvm-dev] How to prevent llvm's default optimization
This is likely more of a canonicalization than an optimization. This is done so that if you got the add followed by mul input or the mul followed by add input they would be canonicalized to the same sequence. Maybe not the optimal sequence but at least the same. I didn't check, but I suspect this is happening in InstCombine in the middle end. ~Craig On Mon, Jun 29, 2020 at 8:37 PM Ben Shi via llvm-dev < llvm-dev at lists.llvm.org> wrote:> Hi, James, > > Thanks for your reply. > > I do not think it is always true, that "mul then add" is faster than "add > then mul". > > For example, > > A small immediate can be directly encoded in the instruction, but it > becomes a larger one after a multiplication, which has to be loaded from > the constant pool (extra memory access). > > So I wonder, is it possile to prevent it, via changes to the config of the > base class TargetLowering,than writing special custom C++ code. > > > Ben > > > > > > > > At 2020-06-30 07:46:41, "James Courtier-Dutton" <james.dutton at gmail.com> wrote: > >Hi Ben, > > > >Why do you want to stop it? > >"mul then add" is faster than "add then mul". > >The result is the same in both cases. > > > >On Mon, 29 Jun 2020 at 22:11, Ben Shi via llvm-dev > ><llvm-dev at lists.llvm.org> wrote: > >> > >> Hello, > >> I have an instruction pattern like > >> %2 = add i32 %0, 25 > >> %3 = mul i32 %2, 525 > >> > >> and llvm will optimize it to > >> %2 = mul i32 %0, 525 > >> %3 = add i32 %2, 12525 > >> > >> how to prevent it? > >> > >> _______________________________________________ > >> LLVM Developers mailing list > >> llvm-dev at lists.llvm.org > >> https://lists.llvm.org/cgi-bin/mailman/listinfo/llvm-dev > > _______________________________________________ > LLVM Developers mailing list > llvm-dev at lists.llvm.org > https://lists.llvm.org/cgi-bin/mailman/listinfo/llvm-dev >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.llvm.org/pipermail/llvm-dev/attachments/20200629/f312c917/attachment.html>
Sanjay Patel via llvm-dev
2020-Jun-30 13:08 UTC
[llvm-dev] How to prevent llvm's default optimization
Yes - this has been in InstCombine for a long time: https://github.com/llvm/llvm-project/blob/master/llvm/lib/Transforms/InstCombine/InstCombineMulDivRem.cpp#L268 We could say that the canonicalization should be reversed, but that probably uncovers more missing optimizations. The code size concern is legitimate. For example on x86, gcc asm is 2 bytes smaller on this example: https://godbolt.org/z/GK9FEL To improve this, we could add a generic transform to DAGCombiner to invert the transform that was done in IR. That transform would only be enabled with a TargetLowering hook that allows targets to decide if the constants or other factors (optimizing for size) make it worthwhile to reorder the ops. On Mon, Jun 29, 2020 at 11:54 PM Craig Topper via llvm-dev < llvm-dev at lists.llvm.org> wrote:> This is likely more of a canonicalization than an optimization. This is > done so that if you got the add followed by mul input or the mul followed > by add input they would be canonicalized to the same sequence. Maybe not > the optimal sequence but at least the same. I didn't check, but I suspect > this is happening in InstCombine in the middle end. > > ~Craig > > > On Mon, Jun 29, 2020 at 8:37 PM Ben Shi via llvm-dev < > llvm-dev at lists.llvm.org> wrote: > >> Hi, James, >> >> Thanks for your reply. >> >> I do not think it is always true, that "mul then add" is faster than >> "add then mul". >> >> For example, >> >> A small immediate can be directly encoded in the instruction, but it >> becomes a larger one after a multiplication, which has to be loaded from >> the constant pool (extra memory access). >> >> So I wonder, is it possile to prevent it, via changes to the config of >> the base class TargetLowering,than writing special custom C++ code. >> >> >> Ben >> >> >> >> >> >> >> >> At 2020-06-30 07:46:41, "James Courtier-Dutton" <james.dutton at gmail.com> wrote: >> >Hi Ben, >> > >> >Why do you want to stop it? >> >"mul then add" is faster than "add then mul". >> >The result is the same in both cases. >> > >> >On Mon, 29 Jun 2020 at 22:11, Ben Shi via llvm-dev >> ><llvm-dev at lists.llvm.org> wrote: >> >> >> >> Hello, >> >> I have an instruction pattern like >> >> %2 = add i32 %0, 25 >> >> %3 = mul i32 %2, 525 >> >> >> >> and llvm will optimize it to >> >> %2 = mul i32 %0, 525 >> >> %3 = add i32 %2, 12525 >> >> >> >> how to prevent it? >> >> >> >> _______________________________________________ >> >> LLVM Developers mailing list >> >> llvm-dev at lists.llvm.org >> >> https://lists.llvm.org/cgi-bin/mailman/listinfo/llvm-dev >> >> _______________________________________________ >> LLVM Developers mailing list >> llvm-dev at lists.llvm.org >> https://lists.llvm.org/cgi-bin/mailman/listinfo/llvm-dev >> > _______________________________________________ > LLVM Developers mailing list > llvm-dev at lists.llvm.org > https://lists.llvm.org/cgi-bin/mailman/listinfo/llvm-dev >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.llvm.org/pipermail/llvm-dev/attachments/20200630/25c5ad68/attachment.html>