similar to: CUDA tools?

Displaying 20 results from an estimated 30000 matches similar to: "CUDA tools?"

2017 Oct 05
4
CUDA tools?
vychytraly . wrote: > On Thu, Oct 5, 2017 at 9:51 PM, <m.roth at 5-cent.us> wrote: >> >> So, kmod-nvidia installed. Trouble is, I have no tool to test it. And my >> user might need nvcc, which, of course, is only provided by the NVidia >> CUDA, which won't install, because it conflicts with kmod-nvidia. >> >> Has *anyone* dealt with this? If so,
2017 Oct 06
0
CUDA tools?
On Thu, 2017-10-05 at 17:07 -0400, m.roth at 5-cent.us wrote: > vychytraly . wrote: > > On Thu, Oct 5, 2017 at 9:51 PM, <m.roth at 5-cent.us> wrote: > > > > > > So, kmod-nvidia installed. Trouble is, I have no tool to test it. And my > > > user might need nvcc, which, of course, is only provided by the NVidia > > > CUDA, which won't install,
2008 Nov 04
1
Help needed using 3rd party C library/functions from within R (Nvidia CUDA)
Hello, I'm trying to combine the parallel computing power available through NVIDIA CUDA (www.nvidia.com/cuda) from within R. CUDA is an extension to the C language, so I thought it would be possible to do this. If I have a C file with an empty function which includes a needed CUDA library (cutil.h) and compile this to an .so file using a NVIDIA compiler (nvcc), called 'myFunc.so' I
2016 Oct 27
3
problem on compiling cuda program with clang++
Hi all, I compiled the *llvm3.9* source code on the *Nvidia TX1* board. And now I am following the document in the docs/CompileCudaWithLLVM.rst to compile cuda program with clang++. However, when I compile `axpy.cu` using `nvcc`, *nvcc* can generate the correct the binary; while compiling `axpy.cu` using clang++, the detailed command is `clang++ axpy.cu -o axpy --cuda-gpu-arch=sm_53
2016 Oct 14
2
LLVM/CLANG: CUDA compilation fail for inline assembly code
Hi, I am sorry for sending this query again here, but maybe I sent it to wrong list yesterday. I am trying to compile LonestarGPU-rev2.0 <http://iss.ices.utexas.edu/?p=projects/galois/lonestargpu/download> benchmark suite with LLVM/CLANG. This suite has a following piece of code (more info here
2017 Oct 05
0
CUDA tools?
Are you installing CUDA from official NVidia repository? On Thu, Oct 5, 2017 at 9:51 PM, <m.roth at 5-cent.us> wrote: > Hi, again. > > So, kmod-nvidia installed. Trouble is, I have no tool to test it. And my > user might need nvcc, which, of course, is only provided by the NVidia > CUDA, which won't install, because it conflicts with kmod-nvidia. > > Has *anyone*
2015 Apr 08
2
[LLVMdev] CUDA front-end (CUDA to LLVM IR)
On Wed, Apr 8, 2015 at 10:12 AM, Dmitry Mikushin <dmitry at kernelgen.org> wrote: > A tool of this kind here: https://github.com/apc-llc/nvcc-llvm-ir > > 2015-04-08 19:01 GMT+02:00 Ahmed ElTantawy <ahmede at ece.ubc.ca>: > >> Hi, >> >> I wanted to ask whether there is ongoing effort (or an already >> established tool) that enables to convert CUDA
2015 Aug 21
2
[CUDA/NVPTX] is inlining __syncthreads allowed?
I'm using 7.0. I am attaching the reduced example. nvcc sync.cu -arch=sm_35 -ptx gives // .globl _Z3foov .visible .entry _Z3foov( ) { .reg .pred %p<2>; .reg .s32 %r<3>; mov.u32 %r1, %tid.x; and.b32 %r2, %r1, 1; setp.eq.b32 %p1, %r2, 1; @!%p1 bra BB7_2; bra.uni
2017 Sep 27
2
Semi-OT: hardware: NVidia proprietary driver, C7.4
Phil Perry wrote: > On 27/09/17 16:49, m.roth at 5-cent.us wrote: >> Hi, folks, >> >> Well, still more fun (for values of fun approaching zero): >> >> 1. Went to install CUDA 9.0... well, gee, there is *no* CUDA 9.0. >> Even though I installed the 9 repo, all that I get is 8. I've >> used their webform, and an waiting on a
2011 Aug 15
2
[LLVMdev] Cuda programs on LLVM
Hello , How to execute a cuda program using llvm? More specifically, nvcc produces some temporary files during its compilation. I want to convert the .cu.cpp to .ll format and optimize it. The .cu.cpp file contains typedefs and enums used by cuda runtime and also the host part of the code and the ptx file contains the kernel definition. How can i run the program after optimization? Will Rhodin
2017 Sep 27
2
OrcJIT + CUDA Prototype for Cling
Dear LLVM-Developers and Vinod Grover, we are trying to extend the cling C++ interpreter (https://github.com/root-project/cling) with CUDA functionality for Nvidia GPUs. I already developed a prototype based on OrcJIT and am seeking for feedback. I am currently a stuck with a runtime issue, on which my interpreter prototype fails to execute kernels with a CUDA runtime error. === How to use the
2017 Sep 27
3
Semi-OT: hardware: NVidia proprietary driver, C7.4
Hi, folks, Well, still more fun (for values of fun approaching zero): 1. Went to install CUDA 9.0... well, gee, there is *no* CUDA 9.0. Even though I installed the 9 repo, all that I get is 8. I've used their webform, and an waiting on a reply. 2. I remove all nvidia packages. 3. It appears that the kmod-nvidia is what I need; that's what nvidia-detect
2012 Nov 09
0
[LLVMdev] translating from OpenMP to CUDA
The PTX back-end is robust (it's based on the sources used by nvcc), but I'm not sure about the OpenMP representation in LLVM IR. I believe the OpenMP constructs are already lowered into libgomp calls before leaving DragonEgg. It's been awhile since I've loooked at it though. If you use the PTX back-end and have any issues, please don't hesitate to post to the list and cc:
2011 Aug 15
0
[LLVMdev] Cuda programs on LLVM
Hi Adarsh, to my knowledge there is no publicly available CUDA-Frontend for LLVM yet. The work of Helge Rhodin you mentioned is on the backend-side: It allows to generate PTX code from LLVM IR. It is still being maintained, although I think the currently available source code is a little outdated. There is also a PTX backend in the current version of LLVM that makes use of LLVM's
2016 Oct 27
0
problem on compiling cuda program with clang++
Hi, it looks like you're compiling CUDA for an ARM host? This is not a configuration we have tested, nor is it something we have the capability of testing at the moment. You may be able to make it work by providing the appropriate -isystem flags to clang so that it can find your headers, but who knows, it may be more complicated than that. Regards, -Justin On Wed, Oct 26, 2016 at 9:59 PM,
2018 Feb 20
2
use clang++ to build lulesh 2.0 failed
Hello, I'm trying to use clang++ instead of nvcc to build lulesh 2.0 cuda version. And it fails in the compilation with errors like the one below: opt/common/cuda/cuda-7.5.18/include/thrust/iterator/iterator_adaptor.h:187:5: error: expected member name or ';' after declaration specifiers __thrust_exec_check_disable__ ​It looks like clang++ is complaining about the thrust library
2018 Feb 20
0
use clang++ to build lulesh 2.0 failed
> ​It looks like clang++ is complaining about the thrust library comes with cuda, The Thrust library that comes with CUDA is indeed not compatible with clang. We made a number of changes to Thrust to make it work with clang (it was relying on what we considered to be bugs in nvcc), but they're only available in the upstream Thrust: https://github.com/thrust/thrust. No promises that one
2015 Apr 08
5
[LLVMdev] CUDA front-end (CUDA to LLVM IR)
Hi, I wanted to ask whether there is ongoing effort (or an already established tool) that enables to convert CUDA kernels (that uses CUDA specific intrinsics, e.g., threadId.x, __syncthreads(), ...) to LLVM IR. I am aware that I can do this for OpenCL with the help of libclc but I can not find something similar for CUDA. Thanks -------------- next part -------------- An HTML attachment was
2014 Mar 25
3
NVidia, again
Got a HBS (y'know, Honkin' Big Server, one o' them technical terms), a Dell 720 with two Tesla GPUs. I updated the o/s, 6.5, and I cannot get the GPUs recognized. As a last resort, I d/l NVidia's proprietary driver/installer, 325, and it builds fine... I've yum removed the kmod-nvidia I had on the system, nouveau is blacklisted, and when I reboot, lsmod shows me nvidia loaded,
2017 Jun 09
1
NVPTX Back-end: relocatable device code support for dynamic parallelism
Hi everyone, CUDA allows to call some runtime functions also from the device code. On a multi-GPU system this allows the GPU to determine its device id on its own via cudaGetDevice(). Unfortunately i cannot get it working when compiling with clang. When compiling with nvcc relocatable device code needs to be set to true (-rdc=true) and the cudadevrt is needed when linking [0]. I did not