Mehdi Amini via llvm-dev
2015-Sep-04 16:28 UTC
[llvm-dev] Testing "normal" cross-compilers versus GPU backends
> On Sep 3, 2015, at 5:56 PM, Robinson, Paul <Paul_Robinson at playstation.sony.com> wrote: > > > >> -----Original Message----- >> From: Mehdi Amini [mailto:mehdi.amini at apple.com] >> Sent: Thursday, September 03, 2015 3:26 PM >> To: Robinson, Paul >> Cc: Tom Stellard; llvm-dev at lists.llvm.org; NAKAMURA Takumi >> Subject: Re: Testing "normal" cross-compilers versus GPU backends >> >> >>> On Sep 3, 2015, at 11:23 AM, Robinson, Paul >> <Paul_Robinson at playstation.sony.com> wrote: >>> >>> >>> >>>> -----Original Message----- >>>> From: Tom Stellard [mailto:tom at stellard.net] >>>> Sent: Thursday, September 03, 2015 7:31 AM >>>> To: Mehdi Amini >>>> Cc: Robinson, Paul; llvm-dev at lists.llvm.org; NAKAMURA Takumi >>>> Subject: Re: Testing "normal" cross-compilers versus GPU backends >>>> >>>> On Thu, Sep 03, 2015 at 02:07:54AM -0700, Mehdi Amini wrote: >>>>> >>>>>> On Sep 3, 2015, at 12:18 AM, Robinson, Paul >>>> <Paul_Robinson at playstation.sony.com> wrote: >>>>>> >>>>>> >>>>>> >>>>>>> -----Original Message----- >>>>>>> From: Mehdi Amini [mailto:mehdi.amini at apple.com] >>>>>>> Sent: Wednesday, September 02, 2015 7:10 PM >>>>>>> To: Robinson, Paul >>>>>>> Cc: llvm-dev at lists.llvm.org; tom at stellard.net; NAKAMURA Takumi >>>>>>> Subject: Re: Testing "normal" cross-compilers versus GPU backends >>>>>>> >>>>>>> Hi Paul, >>>>>>> >>>>>>> Thanks for the summary! >>>>>>> >>>>>>>> On Sep 2, 2015, at 5:44 PM, Robinson, Paul >>>>>>> <Paul_Robinson at playstation.sony.com> wrote: >>>>>>>> >>>>>>>> This note arose from >>>> https://urldefense.proofpoint.com/v2/url?u=http- >>>>>>> 3A__reviews.llvm.org_D12506&d=BQIFAg&c=eEvniauFctOgLOKGJOplqw&r=v- >>>>>>> >>>> >> ruWq0KCv2O3thJZiK6naxuXK8mQHZUmGq5FBtAmZ4&m=Wr0uOhkAp_10X4edWwxZQ9V8L97j8e >>>>>>> o6cR_1Ia-gMOw&s=OOTP9DnL-TWV1zvy9EcU0Z6yfTq5lBjhE-LvYlWMJ3Y&e= but >>>> the >>>>>>> reviewers >>>>>>>> felt that we needed a broader audience, because the proposed patch >>>>>>>> really didn't solve the entire problem and we had no better ideas. >>>>>>>> >>>>>>>> Mehdi Amini needs to build LLVM with just a GPU backend, and still >>>> have >>>>>>>> "ninja check" Just Work. Commits r243958-243960 tried to >> accomplish >>>>>>>> that; however they are too big a hammer, and cause much simpler >>>> cross >>>>>>>> environments (like mine) to exclude big chunks of very useful tests >>>>>>>> (including my favorite, DebugInfo). >>>>>>>> >>>>>>>> FYI, my main cross environment is building on X86 Windows but using >>>>>>>> a default target triple for PS4 (which is also X86). >>>>>>>> >>>>>>>> I experimented with building LLVM with just the ARM backend >> (running >>>> on >>>>>>>> an X86 workstation) and setting the default triple to some ARM >>>> value. >>>>>>>> "ninja check" worked fine (without Mehdi's series of commits), so >>>> the >>>>>>>> normal kind of cross-compiler environment seems to be happy with >> how >>>>>>>> things were set up originally. >>>>>>>> >>>>>>>> Mehdi reports building LLVM with the X86 and AMDGPU backends, >>>> setting >>>>>>>> the default triple to "amdgcn--amdhsa", and getting 200-some >>>> failures. >>>>>>>> >>>>>>>> (This does make me wonder about AMDGPU testing in general; how does >>>> that >>>>>>>> work? The only places I see lit checks for AMDGPU are in the usual >>>>>>>> target-dependent places.) >>>>>>> >>>>>>> I don’t understand this interrogation about how do you do testing in >>>>>>> general. The same way you don’t process tests/CodeGen/X86/* with the >>>> ARM >>>>>>> backend, you can’t process any random IR through these backends. >>>>>> >>>>>> You said you had 200+ failures with AMDGPU. Are the AMD folks simply >>>>>> tolerating the 200 failures, and you don't want to? I should hope >>>> there >>>>>> is more to it than that. >>>>> >>>>> Well, I don’t know, they might just run `ninja check` with the default >>>> triple set to X86? >>>>> (which I would consider being working around a buggy test suite) >>>>> >>>> >>>> I always enable AMDGPU and X86 when I build, so I've never run into >>>> this problem. >>>> >>>> -Tom >>> >>> Tom, presumably your default target triple is X86-ish? And the only >>> tests to exercise the AMDGPU backend are those that explicitly specify >>> a triple for AMDGPU? >> >> This is how we used to do it as well, and I assume this is how most of >> backends out of the main CPUs are dealing with the situation. >> This is what I consider a bug and was trying to solve (admittedly not in >> an optimal way). >> >>> Mehdi, assuming that's what Tom does, your stated goal was to be able to >>> run tests *without* including the X86 backend, so Tom's solution won't >>> work for you (restating just for confirmation). >> >> Yes, or alternatively I expect the test suite to pass whatever the default >> triple is. > > Um, that may be asking a bit much…It already works for >95% of the suite :) The rest is working (and useful) only for “some” (unspecified) cross compiler configuration.> >> >>> Krzysztof suggested much the same thing that I think you are currently >>> doing, which is deliberately configure a default triple but exclude the >>> corresponding backend. >> >> You and Takumi were considering this as an unsupported configuration >> before, and I tend to agree with that (this is the configuration I’m using >> for our tests but it was not intentional to leave the default triple >> unset). > > Right, intuitively it doesn't make sense. Is it actually useful to build a > GPU compiler that will crash unless you ask it to generate GPU code? Seems > to me it should default to producing GPU code.Correct me if I’m wrong: You’re viewing this from the “clang” point of view. A default triple is needed because the command line interface does not require to specify it. I see LLVM as a library or compiler *framework* in the first place, and clang is just a use case as another. When you build a compiler using LLVM as a library: 1) it does not have to be a command line compiler, 2) the interface does not have to make optional the target selection Most GPU compilers are embedded in the driver (they compile shaders on-demand during host program execution). The driver can detect the hardware and initialize LLVM with the right triple. We build LLVM as a shared library, we then build multiple compiler that will link to this library to CodeGen to various backend. The compiler is responsible to select and initialize the appropriate backend, we *never* rely on the default triple, and I don’t even see how we could. You could also see LLVM as a system library that can have multiple clients, each client responsible of its own initialization.>>> I expect we can detect that situation in lit.cfg >>> and exclude tests on that basis, rather than 'native'. It would solve >>> the problem for my case (host triple != target triple, although the arch >>> parts of the triple do match) and the "normal" cross-compiler case (e.g. >>> host = X86, backend + target triple = ARM). >>> >>> I'm going to play around with that and see what I can do to make it >> work. >>> >>>> >>>> >>>>>> >>>>>>> >>>>>>> IMO, the problem is in general about tests that are written without >>>>>>> specifying a triple, that will be executed with the default triple. >>>>>>> >>>>>>> Most of these tests were written with X86 (or ARM) in mind, and >> there >>>> is >>>>>>> no guarantee that they will behave as intended with every possible >>>> triple. >>>>>>> The DataLayout for instance has to be the one from the target, and >> is >>>> not >>>>>>> portable. >>>>>>> I think a "portable backend test” is pretty rare in general. >>>>>> >>>>>> It depends on what the test is trying to do. I'm sure it is quite >>>> common >>>>>> for IR tests to behave essentially the same way regardless of target. >>>>> >>>>> IR tests != backend test (I may miss your point here, it’s late…). >>> >>> Right, sorry, lost focus for a moment there... nevertheless it is still >>> the case that many tests exercise functionality that is not particularly >>> target-centric and these should be run for any target that actually >>> supports that functionality. For example, the DebugInfo tests should >>> be run for any target that supports emitting debug info. >> >> I’m not sure that “debug info” support is all or nothing. >> As an extreme example, I know targets that support debug info but do not >> support function calls, what if your "debug info” test involves these? > > Then as part of getting the test suite to work for you, you would need to > disable that particular test for your target. It sounds like this kind of > thing is exactly what the Hexagon folks did, and it seems quite reasonable. > (And in fact I see two DebugInfo tests marked XFAIL: hexagon.)It seems conceptually wrong to me, for the reason I already exposed. It should go the other way (whitelist instead of blacklist)> >> >> Also, I’m not a DebugInfo expert, but when a front-end generated them, >> aren’t they dependent on the DataLayout? Hence the target? > > Not really. DebugInfo tests primarily care what the DWARF description > looks like, not so much what the generated code looks like,My question is less will the CHECK matches than “will the backend be able to generate code with invalid debug information (ex: pointer size, etc.) or just crash?”> and there > are 100 or so DebugInfo tests that work across lots of targets. > > Many DebugInfo tests are in target-dependent directories as artifacts of > how the tests are implemented, rather than because they are truly > target-dependent. The ones I've tripped over tend to be target-specific > because they generate assembler and the assembler syntax varies too much > across targets to be worth making the CHECKs tolerant enough. > > There are fine details that would be DataLayout dependent, but 99% of > the DebugInfo tests aren't checking things that are at such a low level. > >> It means that many DebugInfo test could fail with a backend that would >> have a significantly different DataLayout and won’t expect some stuff the >> way they are played down. >> >> >>> Whether a >>> target supports debug info is orthogonal to its native-ness. As written >>> the top-level DebugInfo tests should work fine as long as the default >>> triple's backend is included, and that backend supports debug info. >>> >>> If your backend doesn't support debug info, then it's reasonable to >>> exclude those from your testing; but we can't do that at the cost of >>> excluding those from testing other backends that *do* support the >> feature, >>> even if that testing runs in a cross-compiler environment. >>> >>> In this particular example, we'd be setting things up so that DebugInfo >>> is excluded for the wrong reason (not based on some more abstract idea >>> of the feature-set of the target) but as Krzysztof observes, getting >>> a feature-oriented set of conditions would be a huge task. >> >> Agree: the predicate I wrote is not correct, and I don’t see a trivial >> “good” solution. This is why any compromise or intermediate solution that >> is good enough to support whatever use-case people here have, including >> your, seems appropriate to implement. >> >> Let me know I can help coding something. > > I have experimented with implementing the thing Takumi and I think should > be a configuration error. :-) Basically it takes the same kind of approach > that I did in D12506, except it checks for the existence of the target that > matches the default triple. If that target exists then 'llc' with no triple > will succeed, and it looks like the bulk of the tests that you disabled are > in that category. I'm not especially happy about this tactic, though.Why aren’t you happy about that?> > The Hexagon precedent is interesting; Krzysztof said they set the default > triple, and didn't have to xfail all that much stuff. Searching the tree, > I find exactly 7 individual tests marked XFAIL: hexagon, plus it disables > all of ExecutionEngine, and turns off the 'object-emission' feature. > > I'm curious if you would try setting the default triple to match your > target, and see what /kinds/ of tests fail. The raw number is much less > interesting than in the categories.Failing tests attached, let me know which ones you’d like me to investigate. — Mehdi
Mehdi Amini via llvm-dev
2015-Sep-04 16:29 UTC
[llvm-dev] Testing "normal" cross-compilers versus GPU backends
> On Sep 4, 2015, at 9:28 AM, Mehdi Amini <mehdi.amini at apple.com> wrote: > >> >> On Sep 3, 2015, at 5:56 PM, Robinson, Paul <Paul_Robinson at playstation.sony.com> wrote: >> >> >> >>> -----Original Message----- >>> From: Mehdi Amini [mailto:mehdi.amini at apple.com] >>> Sent: Thursday, September 03, 2015 3:26 PM >>> To: Robinson, Paul >>> Cc: Tom Stellard; llvm-dev at lists.llvm.org; NAKAMURA Takumi >>> Subject: Re: Testing "normal" cross-compilers versus GPU backends >>> >>> >>>> On Sep 3, 2015, at 11:23 AM, Robinson, Paul >>> <Paul_Robinson at playstation.sony.com> wrote: >>>> >>>> >>>> >>>>> -----Original Message----- >>>>> From: Tom Stellard [mailto:tom at stellard.net] >>>>> Sent: Thursday, September 03, 2015 7:31 AM >>>>> To: Mehdi Amini >>>>> Cc: Robinson, Paul; llvm-dev at lists.llvm.org; NAKAMURA Takumi >>>>> Subject: Re: Testing "normal" cross-compilers versus GPU backends >>>>> >>>>> On Thu, Sep 03, 2015 at 02:07:54AM -0700, Mehdi Amini wrote: >>>>>> >>>>>>> On Sep 3, 2015, at 12:18 AM, Robinson, Paul >>>>> <Paul_Robinson at playstation.sony.com> wrote: >>>>>>> >>>>>>> >>>>>>> >>>>>>>> -----Original Message----- >>>>>>>> From: Mehdi Amini [mailto:mehdi.amini at apple.com] >>>>>>>> Sent: Wednesday, September 02, 2015 7:10 PM >>>>>>>> To: Robinson, Paul >>>>>>>> Cc: llvm-dev at lists.llvm.org; tom at stellard.net; NAKAMURA Takumi >>>>>>>> Subject: Re: Testing "normal" cross-compilers versus GPU backends >>>>>>>> >>>>>>>> Hi Paul, >>>>>>>> >>>>>>>> Thanks for the summary! >>>>>>>> >>>>>>>>> On Sep 2, 2015, at 5:44 PM, Robinson, Paul >>>>>>>> <Paul_Robinson at playstation.sony.com> wrote: >>>>>>>>> >>>>>>>>> This note arose from >>>>> https://urldefense.proofpoint.com/v2/url?u=http- >>>>>>>> 3A__reviews.llvm.org_D12506&d=BQIFAg&c=eEvniauFctOgLOKGJOplqw&r=v- >>>>>>>> >>>>> >>> ruWq0KCv2O3thJZiK6naxuXK8mQHZUmGq5FBtAmZ4&m=Wr0uOhkAp_10X4edWwxZQ9V8L97j8e >>>>>>>> o6cR_1Ia-gMOw&s=OOTP9DnL-TWV1zvy9EcU0Z6yfTq5lBjhE-LvYlWMJ3Y&e= but >>>>> the >>>>>>>> reviewers >>>>>>>>> felt that we needed a broader audience, because the proposed patch >>>>>>>>> really didn't solve the entire problem and we had no better ideas. >>>>>>>>> >>>>>>>>> Mehdi Amini needs to build LLVM with just a GPU backend, and still >>>>> have >>>>>>>>> "ninja check" Just Work. Commits r243958-243960 tried to >>> accomplish >>>>>>>>> that; however they are too big a hammer, and cause much simpler >>>>> cross >>>>>>>>> environments (like mine) to exclude big chunks of very useful tests >>>>>>>>> (including my favorite, DebugInfo). >>>>>>>>> >>>>>>>>> FYI, my main cross environment is building on X86 Windows but using >>>>>>>>> a default target triple for PS4 (which is also X86). >>>>>>>>> >>>>>>>>> I experimented with building LLVM with just the ARM backend >>> (running >>>>> on >>>>>>>>> an X86 workstation) and setting the default triple to some ARM >>>>> value. >>>>>>>>> "ninja check" worked fine (without Mehdi's series of commits), so >>>>> the >>>>>>>>> normal kind of cross-compiler environment seems to be happy with >>> how >>>>>>>>> things were set up originally. >>>>>>>>> >>>>>>>>> Mehdi reports building LLVM with the X86 and AMDGPU backends, >>>>> setting >>>>>>>>> the default triple to "amdgcn--amdhsa", and getting 200-some >>>>> failures. >>>>>>>>> >>>>>>>>> (This does make me wonder about AMDGPU testing in general; how does >>>>> that >>>>>>>>> work? The only places I see lit checks for AMDGPU are in the usual >>>>>>>>> target-dependent places.) >>>>>>>> >>>>>>>> I don’t understand this interrogation about how do you do testing in >>>>>>>> general. The same way you don’t process tests/CodeGen/X86/* with the >>>>> ARM >>>>>>>> backend, you can’t process any random IR through these backends. >>>>>>> >>>>>>> You said you had 200+ failures with AMDGPU. Are the AMD folks simply >>>>>>> tolerating the 200 failures, and you don't want to? I should hope >>>>> there >>>>>>> is more to it than that. >>>>>> >>>>>> Well, I don’t know, they might just run `ninja check` with the default >>>>> triple set to X86? >>>>>> (which I would consider being working around a buggy test suite) >>>>>> >>>>> >>>>> I always enable AMDGPU and X86 when I build, so I've never run into >>>>> this problem. >>>>> >>>>> -Tom >>>> >>>> Tom, presumably your default target triple is X86-ish? And the only >>>> tests to exercise the AMDGPU backend are those that explicitly specify >>>> a triple for AMDGPU? >>> >>> This is how we used to do it as well, and I assume this is how most of >>> backends out of the main CPUs are dealing with the situation. >>> This is what I consider a bug and was trying to solve (admittedly not in >>> an optimal way). >>> >>>> Mehdi, assuming that's what Tom does, your stated goal was to be able to >>>> run tests *without* including the X86 backend, so Tom's solution won't >>>> work for you (restating just for confirmation). >>> >>> Yes, or alternatively I expect the test suite to pass whatever the default >>> triple is. >> >> Um, that may be asking a bit much… > > It already works for >95% of the suite :) > The rest is working (and useful) only for “some” (unspecified) cross compiler configuration. > >> >>> >>>> Krzysztof suggested much the same thing that I think you are currently >>>> doing, which is deliberately configure a default triple but exclude the >>>> corresponding backend. >>> >>> You and Takumi were considering this as an unsupported configuration >>> before, and I tend to agree with that (this is the configuration I’m using >>> for our tests but it was not intentional to leave the default triple >>> unset). >> >> Right, intuitively it doesn't make sense. Is it actually useful to build a >> GPU compiler that will crash unless you ask it to generate GPU code? Seems >> to me it should default to producing GPU code. > > Correct me if I’m wrong: > > You’re viewing this from the “clang” point of view. A default triple is needed because the command line interface does not require to specify it. > > I see LLVM as a library or compiler *framework* in the first place, and clang is just a use case as another. > > When you build a compiler using LLVM as a library: 1) it does not have to be a command line compiler, 2) the interface does not have to make optional the target selection > > Most GPU compilers are embedded in the driver (they compile shaders on-demand during host program execution). The driver can detect the hardware and initialize LLVM with the right triple. > > We build LLVM as a shared library, we then build multiple compiler that will link to this library to CodeGen to various backend. The compiler is responsible to select and initialize the appropriate backend, we *never* rely on the default triple, and I don’t even see how we could. > > You could also see LLVM as a system library that can have multiple clients, each client responsible of its own initialization. > > > >>>> I expect we can detect that situation in lit.cfg >>>> and exclude tests on that basis, rather than 'native'. It would solve >>>> the problem for my case (host triple != target triple, although the arch >>>> parts of the triple do match) and the "normal" cross-compiler case (e.g. >>>> host = X86, backend + target triple = ARM). >>>> >>>> I'm going to play around with that and see what I can do to make it >>> work. >>>> >>>>> >>>>> >>>>>>> >>>>>>>> >>>>>>>> IMO, the problem is in general about tests that are written without >>>>>>>> specifying a triple, that will be executed with the default triple. >>>>>>>> >>>>>>>> Most of these tests were written with X86 (or ARM) in mind, and >>> there >>>>> is >>>>>>>> no guarantee that they will behave as intended with every possible >>>>> triple. >>>>>>>> The DataLayout for instance has to be the one from the target, and >>> is >>>>> not >>>>>>>> portable. >>>>>>>> I think a "portable backend test” is pretty rare in general. >>>>>>> >>>>>>> It depends on what the test is trying to do. I'm sure it is quite >>>>> common >>>>>>> for IR tests to behave essentially the same way regardless of target. >>>>>> >>>>>> IR tests != backend test (I may miss your point here, it’s late…). >>>> >>>> Right, sorry, lost focus for a moment there... nevertheless it is still >>>> the case that many tests exercise functionality that is not particularly >>>> target-centric and these should be run for any target that actually >>>> supports that functionality. For example, the DebugInfo tests should >>>> be run for any target that supports emitting debug info. >>> >>> I’m not sure that “debug info” support is all or nothing. >>> As an extreme example, I know targets that support debug info but do not >>> support function calls, what if your "debug info” test involves these? >> >> Then as part of getting the test suite to work for you, you would need to >> disable that particular test for your target. It sounds like this kind of >> thing is exactly what the Hexagon folks did, and it seems quite reasonable. >> (And in fact I see two DebugInfo tests marked XFAIL: hexagon.) > > It seems conceptually wrong to me, for the reason I already exposed. > It should go the other way (whitelist instead of blacklist) > >> >>> >>> Also, I’m not a DebugInfo expert, but when a front-end generated them, >>> aren’t they dependent on the DataLayout? Hence the target? >> >> Not really. DebugInfo tests primarily care what the DWARF description >> looks like, not so much what the generated code looks like, > > My question is less will the CHECK matches than “will the backend be able to generate code with invalid debug information (ex: pointer size, etc.) or just crash?” > > >> and there >> are 100 or so DebugInfo tests that work across lots of targets. >> >> Many DebugInfo tests are in target-dependent directories as artifacts of >> how the tests are implemented, rather than because they are truly >> target-dependent. The ones I've tripped over tend to be target-specific >> because they generate assembler and the assembler syntax varies too much >> across targets to be worth making the CHECKs tolerant enough. >> >> There are fine details that would be DataLayout dependent, but 99% of >> the DebugInfo tests aren't checking things that are at such a low level. >> >>> It means that many DebugInfo test could fail with a backend that would >>> have a significantly different DataLayout and won’t expect some stuff the >>> way they are played down. >>> >>> >>>> Whether a >>>> target supports debug info is orthogonal to its native-ness. As written >>>> the top-level DebugInfo tests should work fine as long as the default >>>> triple's backend is included, and that backend supports debug info. >>>> >>>> If your backend doesn't support debug info, then it's reasonable to >>>> exclude those from your testing; but we can't do that at the cost of >>>> excluding those from testing other backends that *do* support the >>> feature, >>>> even if that testing runs in a cross-compiler environment. >>>> >>>> In this particular example, we'd be setting things up so that DebugInfo >>>> is excluded for the wrong reason (not based on some more abstract idea >>>> of the feature-set of the target) but as Krzysztof observes, getting >>>> a feature-oriented set of conditions would be a huge task. >>> >>> Agree: the predicate I wrote is not correct, and I don’t see a trivial >>> “good” solution. This is why any compromise or intermediate solution that >>> is good enough to support whatever use-case people here have, including >>> your, seems appropriate to implement. >>> >>> Let me know I can help coding something. >> >> I have experimented with implementing the thing Takumi and I think should >> be a configuration error. :-) Basically it takes the same kind of approach >> that I did in D12506, except it checks for the existence of the target that >> matches the default triple. If that target exists then 'llc' with no triple >> will succeed, and it looks like the bulk of the tests that you disabled are >> in that category. I'm not especially happy about this tactic, though. > > Why aren’t you happy about that? > >> >> The Hexagon precedent is interesting; Krzysztof said they set the default >> triple, and didn't have to xfail all that much stuff. Searching the tree, >> I find exactly 7 individual tests marked XFAIL: hexagon, plus it disables >> all of ExecutionEngine, and turns off the 'object-emission' feature. >> >> I'm curious if you would try setting the default triple to match your >> target, and see what /kinds/ of tests fail. The raw number is much less >> interesting than in the categories. > > Failing tests attached, let me know which ones you’d like me to investigate.Tests: -------------- next part -------------- A non-text attachment was scrubbed... Name: failing_tests Type: application/octet-stream Size: 12067 bytes Desc: not available URL: <http://lists.llvm.org/pipermail/llvm-dev/attachments/20150904/5e9bc7db/attachment.obj> -------------- next part -------------- (note I forced enable the “native” feature on this run) — Mehdi
James Y Knight via llvm-dev
2015-Sep-04 17:12 UTC
[llvm-dev] Testing "normal" cross-compilers versus GPU backends
>>> The Hexagon precedent is interesting; Krzysztof said they set the default >>> triple, and didn't have to xfail all that much stuff. Searching the tree, >>> I find exactly 7 individual tests marked XFAIL: hexagon, plus it disables >>> all of ExecutionEngine, and turns off the 'object-emission' feature. >>> >>> I'm curious if you would try setting the default triple to match your >>> target, and see what /kinds/ of tests fail. The raw number is much less >>> interesting than in the categories. >> >> Failing tests attached, let me know which ones you’d like me to investigate. > > Tests: > <failing_tests> > > (note I forced enable the “native” feature on this run)So, just looking at the first one in that list for an example, running: llc -mtriple amdgcn test/CodeGen/Generic/2002-04-14-UnexpectedUnsignedType.ll you get an error of: error: unsupported call to function bar in foo ...because apparently AMDGPU simply doesn't support function calls. At all. That is a rather unusual feature to be missing, and I'd imagine is causing a fair number of the other nominally-generic tests to fail too...
Robinson, Paul via llvm-dev
2015-Sep-04 17:19 UTC
[llvm-dev] Testing "normal" cross-compilers versus GPU backends
> >>> Krzysztof suggested much the same thing that I think you are currently > >>> doing, which is deliberately configure a default triple but exclude > the > >>> corresponding backend. > >> > >> You and Takumi were considering this as an unsupported configuration > >> before, and I tend to agree with that (this is the configuration I’m > using > >> for our tests but it was not intentional to leave the default triple > >> unset). > > > > Right, intuitively it doesn't make sense. Is it actually useful to > build a > > GPU compiler that will crash unless you ask it to generate GPU code? > Seems > > to me it should default to producing GPU code. > > Correct me if I’m wrong: > > You’re viewing this from the “clang” point of view. A default triple is > needed because the command line interface does not require to specify it. > > I see LLVM as a library or compiler *framework* in the first place, and > clang is just a use case as another. > > When you build a compiler using LLVM as a library: 1) it does not have to > be a command line compiler, 2) the interface does not have to make > optional the target selection > > Most GPU compilers are embedded in the driver (they compile shaders on- > demand during host program execution). The driver can detect the hardware > and initialize LLVM with the right triple. > > We build LLVM as a shared library, we then build multiple compiler that > will link to this library to CodeGen to various backend. The compiler is > responsible to select and initialize the appropriate backend, we *never* > rely on the default triple, and I don’t even see how we could. > > You could also see LLVM as a system library that can have multiple > clients, each client responsible of its own initialization.If you want to write your tests as unit tests linked against the library, I have no problem with that. If you want to take advantage of the existing tests, the bulk of them are written using command-line tools, some of which make use of the default triple. So, you need to configure your builds appropriately. It's not about clang, it's about the command-line tools used to implement the tests. If you don't like how the tests work, you don't have to use them.> > > > >>> I expect we can detect that situation in lit.cfg > >>> and exclude tests on that basis, rather than 'native'. It would solve > >>> the problem for my case (host triple != target triple, although the > arch > >>> parts of the triple do match) and the "normal" cross-compiler case > (e.g. > >>> host = X86, backend + target triple = ARM). > >>> > >>> I'm going to play around with that and see what I can do to make it > >> work. > >>> > >>>> > >>>> > >>>>>> > >>>>>>> > >>>>>>> IMO, the problem is in general about tests that are written > without > >>>>>>> specifying a triple, that will be executed with the default > triple. > >>>>>>> > >>>>>>> Most of these tests were written with X86 (or ARM) in mind, and > >> there > >>>> is > >>>>>>> no guarantee that they will behave as intended with every possible > >>>> triple. > >>>>>>> The DataLayout for instance has to be the one from the target, and > >> is > >>>> not > >>>>>>> portable. > >>>>>>> I think a "portable backend test” is pretty rare in general. > >>>>>> > >>>>>> It depends on what the test is trying to do. I'm sure it is quite > >>>> common > >>>>>> for IR tests to behave essentially the same way regardless of > target. > >>>>> > >>>>> IR tests != backend test (I may miss your point here, it’s late…). > >>> > >>> Right, sorry, lost focus for a moment there... nevertheless it is > still > >>> the case that many tests exercise functionality that is not > particularly > >>> target-centric and these should be run for any target that actually > >>> supports that functionality. For example, the DebugInfo tests should > >>> be run for any target that supports emitting debug info. > >> > >> I’m not sure that “debug info” support is all or nothing. > >> As an extreme example, I know targets that support debug info but do > not > >> support function calls, what if your "debug info” test involves these? > > > > Then as part of getting the test suite to work for you, you would need > to > > disable that particular test for your target. It sounds like this kind > of > > thing is exactly what the Hexagon folks did, and it seems quite > reasonable. > > (And in fact I see two DebugInfo tests marked XFAIL: hexagon.) > > It seems conceptually wrong to me, for the reason I already exposed. > It should go the other way (whitelist instead of blacklist)I think the default expectation for a new target is that it should support basically all target-independent LLVM features, and there are tests for lots of those features. Therefore if a target fails to support something, the owner of the target is responsible for making sure those tests are disabled appropriately. It sounds like you think targets should get to pick and choose which things they want to support, and then have to go do work to explicitly enable the target-independent tests for the things they choose to support. We disagree on this point, and I think historically my view is where the community is coming from; this is just my opinion however and I could easily be wrong, but it is clearly how the test suite operates. If you want to reorganize the test suite along different lines, you should start your own llvm-dev discussion about that.> > > > >> > >> Also, I’m not a DebugInfo expert, but when a front-end generated them, > >> aren’t they dependent on the DataLayout? Hence the target? > > > > Not really. DebugInfo tests primarily care what the DWARF description > > looks like, not so much what the generated code looks like, > > My question is less will the CHECK matches than “will the backend be able > to generate code with invalid debug information (ex: pointer size, etc.) > or just crash?”The target's address size is given by the DataLayout, which all targets are required to define, so I'm completely failing to understand the problem. Why would the debug info be invalid? Offhand I can't think of anything the target has to do specially to support debug info, it should Just Work. The details of the object-file format matter more than the target itself, in my experience, and even that matters relatively little. If your target supports emitting arbitrary data, and assembler labels/object-file relocations, then your debug info should be fine. DebugInfo /tests/ end up "depending" on the target only because there's no convenient way to see what the debug info looks like, without producing either an assembly file or an object file. Producing either of those things requires passing data through the target. The details of the target itself are largely irrelevant (although as I've mentioned before, details of the assembler syntax might matter for writing actual CHECK lines). If your target can't produce an object file, then you should turn off 'object-emission' (Hexagon does this, for example). Other than that, I don't see any reason why debug-info tests can't in principle be target-neutral.> > I have experimented with implementing the thing Takumi and I think > should > > be a configuration error. :-) Basically it takes the same kind of > approach > > that I did in D12506, except it checks for the existence of the target > that > > matches the default triple. If that target exists then 'llc' with no > triple > > will succeed, and it looks like the bulk of the tests that you disabled > are > > in that category. I'm not especially happy about this tactic, though. > > Why aren’t you happy about that?Because it takes what would seem to be a configuration error and makes it something we explicitly tolerate. If people conclude that it should be tolerated, then that's fine.> > > > > The Hexagon precedent is interesting; Krzysztof said they set the > default > > triple, and didn't have to xfail all that much stuff. Searching the > tree, > > I find exactly 7 individual tests marked XFAIL: hexagon, plus it > disables > > all of ExecutionEngine, and turns off the 'object-emission' feature. > > > > I'm curious if you would try setting the default triple to match your > > target, and see what /kinds/ of tests fail. The raw number is much less > > interesting than in the categories. > > Failing tests attached, let me know which ones you’d like me to > investigate. > > — > Mehdi
Mehdi Amini via llvm-dev
2015-Sep-04 18:32 UTC
[llvm-dev] Testing "normal" cross-compilers versus GPU backends
> On Sep 4, 2015, at 10:19 AM, Robinson, Paul <Paul_Robinson at playstation.sony.com> wrote: > >>>>> Krzysztof suggested much the same thing that I think you are currently >>>>> doing, which is deliberately configure a default triple but exclude >> the >>>>> corresponding backend. >>>> >>>> You and Takumi were considering this as an unsupported configuration >>>> before, and I tend to agree with that (this is the configuration I’m >> using >>>> for our tests but it was not intentional to leave the default triple >>>> unset). >>> >>> Right, intuitively it doesn't make sense. Is it actually useful to >> build a >>> GPU compiler that will crash unless you ask it to generate GPU code? >> Seems >>> to me it should default to producing GPU code. >> >> Correct me if I’m wrong: >> >> You’re viewing this from the “clang” point of view. A default triple is >> needed because the command line interface does not require to specify it. >> >> I see LLVM as a library or compiler *framework* in the first place, and >> clang is just a use case as another. >> >> When you build a compiler using LLVM as a library: 1) it does not have to >> be a command line compiler, 2) the interface does not have to make >> optional the target selection >> >> Most GPU compilers are embedded in the driver (they compile shaders on- >> demand during host program execution). The driver can detect the hardware >> and initialize LLVM with the right triple. >> >> We build LLVM as a shared library, we then build multiple compiler that >> will link to this library to CodeGen to various backend. The compiler is >> responsible to select and initialize the appropriate backend, we *never* >> rely on the default triple, and I don’t even see how we could. >> >> You could also see LLVM as a system library that can have multiple >> clients, each client responsible of its own initialization. > > If you want to write your tests as unit tests linked against the library, > I have no problem with that.This is not what it is about. It is about build LLVM and running `make check` and having a test suite that test what I built. My tests don’t have any issue.> > If you want to take advantage of the existing tests, the bulk of them are > written using command-line tools, some of which make use of the default > triple. So, you need to configure your builds appropriately. It's not > about clang, it's about the command-line tools used to implement the tests. > If you don't like how the tests work, you don't have to use them.The tools can use the default triple, they don’t *have* to, and especially test don’t *have to* be written this way. Let’s put it this way: Are you considering that building and shipping LLVM without ARM or X86 built-in is a supported configuration for the LLVM project? If yes then `ninja check` must be able to pass with this configuration.> >> >> >> >>>>> I expect we can detect that situation in lit.cfg >>>>> and exclude tests on that basis, rather than 'native'. It would solve >>>>> the problem for my case (host triple != target triple, although the >> arch >>>>> parts of the triple do match) and the "normal" cross-compiler case >> (e.g. >>>>> host = X86, backend + target triple = ARM). >>>>> >>>>> I'm going to play around with that and see what I can do to make it >>>> work. >>>>> >>>>>> >>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>> IMO, the problem is in general about tests that are written >> without >>>>>>>>> specifying a triple, that will be executed with the default >> triple. >>>>>>>>> >>>>>>>>> Most of these tests were written with X86 (or ARM) in mind, and >>>> there >>>>>> is >>>>>>>>> no guarantee that they will behave as intended with every possible >>>>>> triple. >>>>>>>>> The DataLayout for instance has to be the one from the target, and >>>> is >>>>>> not >>>>>>>>> portable. >>>>>>>>> I think a "portable backend test” is pretty rare in general. >>>>>>>> >>>>>>>> It depends on what the test is trying to do. I'm sure it is quite >>>>>> common >>>>>>>> for IR tests to behave essentially the same way regardless of >> target. >>>>>>> >>>>>>> IR tests != backend test (I may miss your point here, it’s late…). >>>>> >>>>> Right, sorry, lost focus for a moment there... nevertheless it is >> still >>>>> the case that many tests exercise functionality that is not >> particularly >>>>> target-centric and these should be run for any target that actually >>>>> supports that functionality. For example, the DebugInfo tests should >>>>> be run for any target that supports emitting debug info. >>>> >>>> I’m not sure that “debug info” support is all or nothing. >>>> As an extreme example, I know targets that support debug info but do >> not >>>> support function calls, what if your "debug info” test involves these? >>> >>> Then as part of getting the test suite to work for you, you would need >> to >>> disable that particular test for your target. It sounds like this kind >> of >>> thing is exactly what the Hexagon folks did, and it seems quite >> reasonable. >>> (And in fact I see two DebugInfo tests marked XFAIL: hexagon.) >> >> It seems conceptually wrong to me, for the reason I already exposed. >> It should go the other way (whitelist instead of blacklist) > > I think the default expectation for a new target is that it should support > basically all target-independent LLVM features, and there are tests for lots > of those features. Therefore if a target fails to support something, the > owner of the target is responsible for making sure those tests are disabled > appropriately.Is there such a documentation of what is a target independent feature that every backend *must* implement? What if I write a test with types like i3 and i1985? My understanding, I may be wrong of course, is that the IR is generic and IR passes should pass on “any” generic IR (i.e. “opt” level passes). As soon as you introduce a backend, you lose these guarantees (i.e. “llc” level).> It sounds like you think targets should get to pick and choose which things > they want to support, and then have to go do work to explicitly enable the > target-independent tests for the things they choose to support. > > We disagree on this point, and I think historically my view is where the > community is coming from; this is just my opinion however and I could > easily be wrong,> but it is clearly how the test suite operates.This is how a ~1% of the test suite operates. It is far from clear to me that this is an intended and controlled decision, and instead looks more like an uncatched bug that was introduced at some point.> If you > want to reorganize the test suite along different lines, you should start > your own llvm-dev discussion about that.Sure, I’m currently trying to implement something along the line that what James suggested. — Mehdi> >> >>> >>>> >>>> Also, I’m not a DebugInfo expert, but when a front-end generated them, >>>> aren’t they dependent on the DataLayout? Hence the target? >>> >>> Not really. DebugInfo tests primarily care what the DWARF description >>> looks like, not so much what the generated code looks like, >> >> My question is less will the CHECK matches than “will the backend be able >> to generate code with invalid debug information (ex: pointer size, etc.) >> or just crash?” > > The target's address size is given by the DataLayout, which all targets > are required to define, so I'm completely failing to understand the problem. > Why would the debug info be invalid? > > Offhand I can't think of anything the target has to do specially to support > debug info, it should Just Work. The details of the object-file format > matter more than the target itself, in my experience, and even that matters > relatively little. If your target supports emitting arbitrary data, and > assembler labels/object-file relocations, then your debug info should be fine. > > DebugInfo /tests/ end up "depending" on the target only because there's no > convenient way to see what the debug info looks like, without producing > either an assembly file or an object file. Producing either of those things > requires passing data through the target. The details of the target itself > are largely irrelevant (although as I've mentioned before, details of the > assembler syntax might matter for writing actual CHECK lines). > > If your target can't produce an object file, then you should turn off > 'object-emission' (Hexagon does this, for example). Other than that, I don't > see any reason why debug-info tests can't in principle be target-neutral. > >>> I have experimented with implementing the thing Takumi and I think >> should >>> be a configuration error. :-) Basically it takes the same kind of >> approach >>> that I did in D12506, except it checks for the existence of the target >> that >>> matches the default triple. If that target exists then 'llc' with no >> triple >>> will succeed, and it looks like the bulk of the tests that you disabled >> are >>> in that category. I'm not especially happy about this tactic, though. >> >> Why aren’t you happy about that? > > Because it takes what would seem to be a configuration error and makes it > something we explicitly tolerate. If people conclude that it should be > tolerated, then that's fine. > >> >>> >>> The Hexagon precedent is interesting; Krzysztof said they set the >> default >>> triple, and didn't have to xfail all that much stuff. Searching the >> tree, >>> I find exactly 7 individual tests marked XFAIL: hexagon, plus it >> disables >>> all of ExecutionEngine, and turns off the 'object-emission' feature. >>> >>> I'm curious if you would try setting the default triple to match your >>> target, and see what /kinds/ of tests fail. The raw number is much less >>> interesting than in the categories. >> >> Failing tests attached, let me know which ones you’d like me to >> investigate. >> >> — >> Mehdi >
Possibly Parallel Threads
- Testing "normal" cross-compilers versus GPU backends
- Testing "normal" cross-compilers versus GPU backends
- Testing "normal" cross-compilers versus GPU backends
- Testing "normal" cross-compilers versus GPU backends
- Testing "normal" cross-compilers versus GPU backends