similar to: [LLVMdev] GSOC Projects

Displaying 20 results from an estimated 20000 matches similar to: "[LLVMdev] GSOC Projects"

2013 Feb 27
2
[LLVMdev] [cfe-dev] GSOC Projects
Renato, Actually I wanted to send GSOC email about the open projects tonight. So, yes - please add them to the corresponding ideas pages. On Wed, Feb 27, 2013 at 10:57 PM, Chris Lattner <clattner at apple.com> wrote: > > On Feb 27, 2013, at 9:24 AM, Renato Golin <renato.golin at linaro.org> wrote: > >> Hi all, >> >> I have some ideas for GSOC projects using
2013 Feb 27
0
[LLVMdev] [cfe-dev] GSOC Projects
Ok, I've added the LNT statistics project, I think it's better to add the Clang stuff to http://clang.llvm.org/OpenProjects.html right? cheers, --renato On 27 February 2013 20:17, Anton Korobeynikov <anton at korobeynikov.info>wrote: > Renato, > > Actually I wanted to send GSOC email about the open projects tonight. > So, yes - please add them to the corresponding
2013 Feb 27
0
[LLVMdev] GSOC Projects
On Feb 27, 2013, at 9:24 AM, Renato Golin <renato.golin at linaro.org> wrote: > Hi all, > > I have some ideas for GSOC projects using LLVM, where should I post them? Please add them to llvm-www/OpenProjects.html. -Chris > > Idea #1: llvm-env > > A simple tool, with little or no dependency on LLVM itself, that will investigate a target architecture by probing
2013 Feb 27
0
[LLVMdev] GSOC Projects
Here's another project idea: We really have needed a `clang-config` for a long time; it's absence makes building programs depending on Clang a pain. -- Sean Silva
2015 May 10
2
[LLVMdev] http://llvm.org/perf/ instability: some clues
Daniel, Tobias, Renato and myself have been looking a little bit at the potential underlying reason for why http://llvm.org/perf/ is instable, and have found some clues. I want to share them here to give people with more experience in the frameworks used by LNT (flask, sqlalchemy, wsgi, .) a chance to check if our reasoning below seems plausible. Daniel noticed the following backtrace in the
2017 Jul 31
2
[LNT] new server instance http://lnt.llvm.org seems unstable
Hi, The new LNT server instance http://lnt.llvm.org seems to fail in many cases. Any entrance to a 'Run page' (e.g. http://lnt.llvm.org/db_default/v4/nts/62475) and lately also many perf bots result submissions (e.g. http://lab.llvm.org:8014/builders/clang-native-arm-lnt-perf/builds/2262/steps/test-suite/logs/stdio ) fails with: "500 Internal Server Error". Any ideas? Thanks,
2014 Aug 01
11
[LLVMdev] Dev Meeting BOF: Performance Tracking
All, I'm curious to know if anyone is interested in tracking performance (compile-time and/or execution-time) from a community perspective? This is a much loftier goal then just supporting build bots. If so, I'd be happy to propose a BOF at the upcoming Dev Meeting. Chad
2017 Jul 31
1
[LNT] new server instance http://lnt.llvm.org seems unstable
The run page problem were triggered by one of my commits (sorry) and should be mitigated now, see the thread at http://lists.llvm.org/pipermail/llvm-dev/2017-July/115971.html <http://lists.llvm.org/pipermail/llvm-dev/2017-July/115971.html> I don't know about the submission problems, could they just an occasional network problem or are they a common phenomenon? Chris did some
2017 Jul 31
2
Internal server error when submitting LNT to lnt.llvm.org
Dear all, I recently started to see internal server errors when submitting perf results to lnt.llvm.org http://lab.llvm.org:8011/builders/perf-x86_64-penryn-O3-polly-before-vectorizer/builds/1154/steps/lnt.nightly-test/logs/stdio nt /home/grosser/buildslave/perf-x86_64-penryn-O3-polly-before-vectorizer/tests/nt/build/sample-0/report.simple.csv
2017 Aug 28
2
Buildbot can't submit results to LNT server
Hi, I have recently moved the clang-native-arm-lnt-perf bot from the nt producer to the test-suite producer. It seems to be working fine but it doesn't manage to submit the results to http://lnt.llvm.org/submitRun. If you scroll down to the bottom of [1], you can see this error message: 2017-08-28 07:06:32: submitting result to 'http://lnt.llvm.org/submitRun' error: lnt server:
2017 Aug 03
2
[LNT] new server instance http://lnt.llvm.org seems unstable
On Wed, Aug 2, 2017, at 21:55, Tobias Grosser via llvm-dev wrote: > This started since about 1-2 weeks, I guess. I am not sure about it > either, but seems unfortunate. Maybe we run into some kind of timeout? And another one :( http://lab.llvm.org:8011/builders/perf-x86_64-penryn-O3-polly/builds/1986/steps/lnt.nightly-test/logs/stdio 2017-08-03 00:24:44 CRITICAL: Results were not
2017 Aug 28
2
Buildbot can't submit results to LNT server
Great, good to know it's not just a problem with our bot. Thanks! On 28 August 2017 at 16:24, Chris Matthews <cmatthews5 at apple.com> wrote: > Hi Diana, > > I have seen that issue on some other bots too. I will fix it this week. It is something about how MySQL treats character encodings differently than Postgres. We are checking for the value in the table beforehand, and this
2017 Jun 27
5
LNT Server offline
Hi, So, I owe you all an apology. I was totally unaware that the llvm.org server, running at UIUC, was still in operation and hosting an active LNT instance. I was under the assumption the LNT server ran elsewhere. As a result, the ability for the LNT bots to submit run information to the "old" LNT server is now gone. I am very sorry for this oversight. In speaking with Chris
2016 Sep 17
7
Benchmark LNT weird thread behaviour
Hi James/Chris, You guys have done this before, so I'm guessing you can help me understand what's going on. If my buildbot config is: jobs=2, nt_flags=['--cflag', '-mcpu=cortex-a15', '--use-perf', '--threads=1', '--build-threads=4'] It uses -j4 for build, -j2 for running the tests:
2017 Aug 02
2
[LNT] new server instance http://lnt.llvm.org seems unstable
Actually I just remember LNT has a page for viewing the log: http://lnt.llvm.org/log ... OperationalError: (OperationalError) (2006, 'MySQL server has gone away') 'SELECT `NT_Run`.`ID` AS `NT_Run_ID`, `NT_Run`.`MachineID` AS `NT_Run_MachineID`, `NT_Run`.`OrderID` AS `NT_Run_OrderID`, `NT_Run`.`ImportedFrom` AS `NT_Run_ImportedFrom`, `NT_Run`.`StartTime` AS `NT_Run_StartTime`,
2013 Jan 28
3
[LLVMdev] adding perf machines
Is there a reason why existing buildbots are not generating LNT results? On 2013-01-28, at 11:37 AM, David Blaikie wrote: They're just build bots running LNT - check the build bot configuration code in the zorg llvm project repository. You'll probably need to do some work to get a machine quiet enough to have reliable/useful performance results, though On Jan 28, 2013 8:33 AM,
2013 Jan 28
2
[LLVMdev] adding perf machines
Is O3-vectorize redundant now that the loop vectorizer is enabled by default? On 2013-01-28, at 12:25 PM, David Blaikie wrote: > On Mon, Jan 28, 2013 at 8:57 AM, Redmond, Paul <paul.redmond at intel.com> wrote: >> Is there a reason why existing buildbots are not generating LNT results? > > Those running LNT should be/are: > >
2016 Apr 22
2
RFC: LNT/Test-suite support for custom metrics and test parameterization
On 21 Apr 2016, at 17:44, Sergey Yakoushkin <sergey.yakoushkin at gmail.com<mailto:sergey.yakoushkin at gmail.com>> wrote: Hi Kristof, The way we use LNT, we would run different configuration (e.g. -O3 vs -Os) as different "machines" in LNT's model. O2/O3 is indeed bad example. We're also using different machines for Os/O3 - such parameters apply to all
2008 Mar 25
0
[LLVMdev] Proposal for GSoC project for improving llvm-test testsuite
Another suggestion: a web service that would allow you to submit a test build and it would provide you feedback on deviations from known good for that target. I know a lot of people keep two trees around for this purpose, and I'd personally like to ditch the duplicate if I could. -- Chris On Mar 21, 2008, at 9:22 PM, Rajika Kumarasiri wrote: > Hello every body, > Thank you all
2016 Sep 01
3
Benchmarks for LLVM-generated Binaries
Hi, I've lately been wondering where benchmarks for LLVM-generated binaries are hosted, and whether they're tracked over time. I'm asking because I'm thinking of where to put some benchmarks I've written using the open source Google benchmarking library [0] to test certain costs of XRay-instrumented binaries, the XRay runtime, and other related measurements (effect of