Displaying 20 results from an estimated 200 matches similar to: "A stupid question..."
2012 Jul 11
2
[LLVMdev] [NVPTX] llc -march=nvptx64 -mcpu=sm_20 generates invalid zero align for device function params
Hello,
FYI, this is a bug http://llvm.org/bugs/show_bug.cgi?id=13324
When compiling the following code for sm_20, func params are by some reason
given with .align 0, which is invalid. Problem does not occur if compiled
for sm_10.
> cat test.ll
; ModuleID = '__kernelgen_main_module'
target datalayout = "e-p:64:64-i64:64:64-f64:64:64-n1:8:16:32:64"
target triple =
2013 Mar 01
0
[LLVMdev] NVPTX CUDA_ERROR_NO_BINARY_FOR_GPU
Ok, as I said, the most precise way to figure out what's wrong is to emit LLVM IR first (use clang -emit-llvm ...) and check out how it differs from working examples, for instance, nvptx regression tests.
----- Original message -----
> I'm building this with llvm-c, and accessing these intrinsics via calling
> the intrinsic as if it were a function.
>
> class F_SREG<string
2013 Mar 01
0
[LLVMdev] NVPTX CUDA_ERROR_NO_BINARY_FOR_GPU
Hi Timothy,
I'm not sure what you mean by this working for other intrinsics, but
in this case, I think you want the intrinsic name
llvm.nvvm.read.ptx.sreg.tid.x.
For me, this looks like:
%x = call i32 @llvm.nvvm.read.ptx.sreg.tid.x()
Pete
On Fri, Mar 1, 2013 at 11:51 AM, Timothy Baldridge <tbaldridge at gmail.com> wrote:
> I'm building this with llvm-c, and accessing these
2013 Mar 01
4
[LLVMdev] NVPTX CUDA_ERROR_NO_BINARY_FOR_GPU
I'm building this with llvm-c, and accessing these intrinsics via calling
the intrinsic as if it were a function.
class F_SREG<string OpStr, NVPTXRegClass regclassOut, Intrinsic IntOp> :
NVPTXInst<(outs regclassOut:$dst), (ins),
OpStr,
[(set regclassOut:$dst, (IntOp))]>;
def INT_PTX_SREG_TID_X : F_SREG<"mov.u32 \t$dst, %tid.x;",
2013 Mar 01
1
[LLVMdev] NVPTX CUDA_ERROR_NO_BINARY_FOR_GPU
The identifier INT_PTX_SREG_TID_X is the name of an instruction as the
back-end sees it, and has very little to do with the name you should use in
your IR. Your best bet is to look at the include/llvm/IR/IntrinsicsNVVM.td
file and see the definitions for each intrinsic. Then, the name mapping is
just:
int_foo_bar -> llvm.foo.bar()
int_ prefix becomes llvm., and all underscores turn into
2013 Mar 01
2
[LLVMdev] NVPTX CUDA_ERROR_NO_BINARY_FOR_GPU
I've written a compiler that outputs PTX code, the result seems fairly
reasonable, but I'm not sure the intrinsics are getting compiled correctly.
In addition, when I try load the module using CUDA, I get an
error: CUDA_ERROR_NO_BINARY_FOR_GPU. I'm running this on a 2012 MBP with
a 640M GPU.
PTX Code (for a mandelbrot calculation):
//
// Generated by LLVM NVPTX Back-End
//
2013 Mar 01
0
[LLVMdev] NVPTX CUDA_ERROR_NO_BINARY_FOR_GPU
Timothy,
Those calls to compute grid intrinsics are definitely wrong. In ptx code they should end up into reading special registers, rather than function calls. Try to take some working example and figure out the LLVM IR differences between it and the result of your compiler.
- D.
----- Original message -----
> I've written a compiler that outputs PTX code, the result seems fairly
>
2004 Jun 25
1
Hashtables major:minor and prio
I have a question about a few things:
ok when you have parent x:x and handle x:x link x:x flowid and so on. What is the
max values of each of these, also where these or some of these hex numbers? I tried
this on redhat 7.3 so there may have been some updates, but this is one of the tests
I did.
[root@RL1 root]# tc filter add dev eth0 protocol ip parent 1:0 prio 10 u32 ht fffe::
match ip dst
2012 Nov 09
0
[LLVMdev] [NVPTX] llc -march=nvptx64 -mcpu=sm_20 generates invalid zero align for device function params
Dear all,
I'm attaching a patch that should fix the issue mentioned above. It
simply makes the same check seen in the same file for global
variables:
emitPTXAddressSpace(PTy->getAddressSpace(), O);
if (GVar->getAlignment() == 0)
O << " .align " << (int) TD->getPrefTypeAlignment(ETy);
else
O << " .align " <<
2011 Jan 07
2
how to run linear regression models at once
hey, folks,
I have two very simple questions. I am not quite sure if I can do this using
R.
so, I am analyzing a large data frame with thousands of variables. For
example:
Dependent variables: d1, d2, d3 (i.e., there are three dependent variables)
Independent variables: s1, s2, s3, ......s1000 (i.e., there are 1000
independent variables)
now I want to run simple linear regression analyses of
2002 Sep 17
3
[ smbfs ] Does it support ACLS?
Hello there friends & foes,
I've had if with M$ windows! Next week I'll sabotage my shop (how 'bout
password protecting dualboots on WS) and force LINUX down my users' throats
;-) (oh, don't I like this tone... makes me feel Billy G.) Oops, showstopper!
ACL enabled network filesystem, where is it? I can't go with alpha code nfs4,
so is there a patch to the smb
2003 May 21
1
netstat/ipcs inside jail
Hi, i've got this problem with my jail and i'm abolutly lost as in the why
of it.
I previously posted this on comp.unix.bsd.freebsd.misc but i was advised to
send here
I was unable to find help on google :(
To resume quick, when i'm in a jail, netstat doesn't work properly.
Hopefully i have provided sufficient information for anyone willing to help
me :p
First of all, my system :
2002 Sep 12
2
LDAP PDC problems
Hello there,
I'm running 2.2.5 compiled with ACL and LDAP auth. The PDC used to work
flawlessly using smbpasswd but I want to dual boot the workstations to
Linux so I need a centralized LDAP authentication repository.
Once I reinstalled the LDAPized samba I started populating the directory
with the "well-known" identities using the smbldap tools (bear in mind
I'm more of a
2006 Nov 29
2
75 seconds timeout when negotiating with smbd.
Hello all,
After upgrading to 3.0.23c on my FreeBSD 6.1 box Samba stopped working
on any connections. When connecting to smbd, tcp session is established
(SYN, SYN+ACK, ACK) then a client send a NEGOTIATION packet, and after
some timeout (about 10-15 seconds) this tcp session is closed (by the
client) because no response to NEG. packet was received. Then after 75
seconds (counting from the
2010 Jul 28
2
minimap and Leaderheads not showing up in Civilization IV
Civilization IV Beyond the Sword expansion.
I'm using Wine 1.1.44 with the latest intel drivers (2.12.0) and mesa at 7.8.2. OS is GNU/Linux, Gentoo distribution
Both the minimap at the lower left corner and leaderheads don't show up, showing just a black section on the screen instead. Also, some artifacts are showing up in forests and jungles.
This is very different from the results
2005 May 17
7
CUPDS reboot the whole system
Hello,
Before a couple of days ago I started /usr/local/sbin/cupsd manualy from
console. When I press CTRL+C to interrupt a program, the system change
runlevel and going to reboot. This was on FreeBSD V5.3, today I
installed fresh new 5.4 but the problem is the same.
cups-base-1.1.23.0_3
--
Todor Dragnev <todor.dragnev@gmail.com>
2003 Jul 17
1
device troubles after stable update
After updating my kernel from a 4.7-stable to 4.8-stable update, my SB PCI128 is no longer recognized by the pcm driver.
Here is my dmesg:
FreeBSD 4.8-STABLE #0: Thu Jul 17 14:13:52 CDT 2003
root@hautlos.stout.osu-res.okstate.edu:/usr/obj/usr/src/sys/GENERIC
Timecounter "i8254" frequency 1193182 Hz
CPU: AMD Athlon(tm) Processor (800.03-MHz 686-class CPU)
Origin =
2007 Mar 01
1
GLM problems.
Dear R users I'm new in R management and maybe It's a silly question. I'm
working with GLM to obtain predictive models. I have some problesm with the
prediction instruction:
> DatosTotal <- read.csv("Var_perdizcsv.csv", sep =";")
> edvariable <- edit(DatosTotal)
> pre <- predict(rlfinal, DatosTotal, type = 'probs')
Erro en
2017 Oct 11
1
[PATCH v1 01/27] x86/crypto: Adapt assembly for PIE support
Change the assembly code to use only relative references of symbols for the
kernel to be PIE compatible.
Position Independent Executable (PIE) support will allow to extended the
KASLR randomization range below the -2G memory limit.
Signed-off-by: Thomas Garnier <thgarnie at google.com>
---
arch/x86/crypto/aes-x86_64-asm_64.S | 45 ++++++++-----
arch/x86/crypto/aesni-intel_asm.S
2013 Aug 23
1
[Bridge] challenge of year: connect to LAN using wireless-ap over bridge + unmanaged l2tpv3 tunnel + bridge? it's possible?
hi everyone,
so,first it seemed a trivial question to me, but since I could not find
anybody being neither able to answer
this question nor giving a short config example. after a few sleepless
nights and exhausting all the reading
and research. here I am sharing my problem with all of you, in the hope of
some possible solution/sugestion.
or is it that this is impossible??
below my