similar to: NVidia, again

Displaying 20 results from an estimated 5000 matches similar to: "NVidia, again"

2017 Jun 20
2
CentOS 6 and luksOpen
Upgraded a RAID. Copied everything from backup. And then my manager said I had to encrypt the drive. I've done that, and made the filesystem, but I can't mount it. CentOS 6. I have the entry in /etc/crypttab, and a key in /etc/crypt.pw, and the luks UUID in /etc/fstab. I cannot find the command that tells it to create the device in /dev/mapper from the info in /etc/crypttab. Clues for
2017 Jun 20
2
CentOS 6 and luksOpen
Leon Fauster wrote: >> Am 20.06.2017 um 16:53 schrieb m.roth at 5-cent.us: >> >> Upgraded a RAID. Copied everything from backup. >> >> And then my manager said I had to encrypt the drive. >> >> I've done that, and made the filesystem, but I can't mount it. >> >> CentOS 6. >> I have the entry in /etc/crypttab, and a key in
2014 May 28
3
The state of xfs on CentOS 6?
We're looking at getting an HBR (that's a technical term, honkin' big RAID). What I'm considering is, rather than chopping it up into 14TB or 16TB filesystems, of using xfs for really big filesystems. The question that's come up is: what's the state of xfs on CentOS6? I've seen a number of older threads seeing problems with it - has that mostly been resolved? How does
2017 Nov 13
4
Addressing the problem of noisy GPUs under Nouveau
Hello, Some users have been complaining for years about their GPU sounding like a jet engine at take off. Last year, I finally laid my hand on one of these GPUs and have been trying to fix this issue on and off since then. After failing to find anything in the HW, I figured out that the duty cycle set by nvidia's proprietary driver would be way under the expected value. By randomly changing
2017 Sep 27
2
Semi-OT: hardware: NVidia proprietary driver, C7.4
Phil Perry wrote: > On 27/09/17 16:49, m.roth at 5-cent.us wrote: >> Hi, folks, >> >> Well, still more fun (for values of fun approaching zero): >> >> 1. Went to install CUDA 9.0... well, gee, there is *no* CUDA 9.0. >> Even though I installed the 9 repo, all that I get is 8. I've >> used their webform, and an waiting on a
2015 Feb 19
0
Anyone using torque/pbs/munge?
CentOS 6.6 I've got two servers, server1 and hbs (honkin' big server). Both are running munge, and torque... *separately*. My problem is that I've got users who want to be able to submit from server1 to hbs. I see that munged can be pointed to an alternate keyfile... but is there any way to tell qsub what to use? (And yes, I got on the torque users' list, and I'm trying
2017 Aug 11
4
[PATCH] drm/nouveau: Document nouveau support for Tegra in DRIVER_DESC
nouveau supports the Tegra K1 and higher after the SoC-based GPUs converged with the main GeForce GPU families. Signed-off-by: Rhys Kidd <rhyskidd at gmail.com> --- drivers/gpu/drm/nouveau/nouveau_drv.h | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/drivers/gpu/drm/nouveau/nouveau_drv.h b/drivers/gpu/drm/nouveau/nouveau_drv.h index 822fe1d4d35e..fbe42ec0a5f1 100644 ---
2011 Feb 08
3
intervals {nlme} lower CI greater than upper CI !!!????
Hi folks... check this out.. > GLU<-lme(gluc~rt*cd4+sex+age+rf+nadir+pharmac+factor(hcv)+factor(hbs)+ + haartd+hivdur+factor(arv), + random= ~rt|id, na.action=na.omit) > intervals(GLU)$fixed lower est. upper (Intercept) 67.3467070345 7.362307e+01 7.989944e+01 rt *0.0148050160* 6.249304e-02 1.101811e-01 cd4
2017 Jun 20
0
CentOS 6 and luksOpen
> Am 20.06.2017 um 16:53 schrieb m.roth at 5-cent.us: > > Upgraded a RAID. Copied everything from backup. > > And then my manager said I had to encrypt the drive. > > I've done that, and made the filesystem, but I can't mount it. > > CentOS 6. > I have the entry in /etc/crypttab, and a key in /etc/crypt.pw, and the > luks UUID in /etc/fstab. I cannot
2017 Jun 20
0
CentOS 6 and luksOpen
> Am 20.06.2017 um 17:12 schrieb m.roth at 5-cent.us: > > Leon Fauster wrote: >>> Am 20.06.2017 um 16:53 schrieb m.roth at 5-cent.us: >>> >>> Upgraded a RAID. Copied everything from backup. >>> >>> And then my manager said I had to encrypt the drive. >>> >>> I've done that, and made the filesystem, but I can't
2015 Jun 25
2
What are the restrictions around loading indirect constbuf values
Hello, We recently tracked down a bug on Tesla GPUs (i.e. G80-GT218) whereby it appears that instructions like 00000028: b5000409 08000780 add rn f32 $r2 $r2 neg c0[$a1] 00000040: b500060d 08004780 add rn f32 $r3 $r3 neg c0[$a1+0x4] or with nvdisasm: .headerflags @"EF_CUDA_SM12 EF_CUDA_PTX_SM(EF_CUDA_SM12)" /*0000*/ FADD R2, R2, -c[0x0][A1+0x0]; /*
2017 Sep 26
7
Semi-OT: hardware: NVidia proprietary driver, C7.4
This is really frustrating. I've got a server with two K20c Tesla cards. I need to use the proprietary drivers to use the CUDA toolkit. Btw, I had no trouble at all with building for CentOS 7.3 I have what NVidia claims is the correct driver package, a 340 series. It appears to build, but then fails to load. The only error I see is "no such device", which makes no sense to me, esp.
2019 Feb 27
7
nvidia on 7.6
I have elrepo installed. rpm -qa | grep elrepo gives elrepo-release-7.0.3.el7.elrepo.noarch nvidia detect (which installed) tells me kmod-nvidia-340xx when I do yum install kmod-nvidia-340xx it is saying no package kmod-nvidia-340xx available ? What did I miss? Thanks, Jerry
2024 Sep 11
1
[PATCH v2 3/3] drm/nouveau: Add drm_panic support for nv50+
On Wed, Sep 11, 2024 at 10:19?AM Jocelyn Falempe <jfalempe at redhat.com> wrote: > On 06/09/2024 21:36, James Jones wrote: > > Right, there are 3 iterations of block linear tiling actually. NV50 does > > support scanout of block linear surfaces. All block-linear-capable GPUs > > do. The 3 generations are: > > > > NV5x/G8x/GTXXX line: Original block size.
2020 Sep 24
4
[RFC] Documentation: nouveau: Introduce some nouveau documentation
Op 23-09-2020 om 22:36 schreef Karol Herbst: > On Wed, Sep 23, 2020 at 10:39 PM Jeremy Cline <jcline at redhat.com> wrote: >> >> On Wed, Sep 23, 2020 at 09:02:45PM +0200, Karol Herbst wrote: >>> On Fri, Sep 11, 2020 at 6:21 PM Jeremy Cline <jcline at redhat.com> wrote: >> >> <snip> >> >>> yeah, I think overall this file is a good
2012 Mar 09
4
HT (Hyper Threading) on CentOS 5.5
Hi, Do we need HT enabled on CentOS release 5.5 (Final) and Kernel Version :- 2.6.18-194.el5 Please help me understand the pros and cons of having HT enabled or disabled on the Server Dell R 710. cat /proc/cpuinfo -> http://fpaste.org/K2dT/ Do let me know if anyone needs more information. Regards, Kaushal
2013 Dec 02
6
GTX 760 passed through
Hello, I''ve successfully passed a unmodified GTX 760 to Win7 x64. It involved some driver patching on the client side but I''m close to getting the required steps on the server side, too. Are there any legal issues that might arise from releasing a patch? Bob [image: Inline image 1] _______________________________________________ Xen-devel mailing list Xen-devel@lists.xen.org
2019 Jan 11
4
Yum excluding packages I need
Stephen Berg (Code 7309) wrote: > On 1/11/19 10:41 AM, mark wrote: > >> C7, and I did a yum update --disableexcludes=all, and yet it's telling >> me [nvidia]: excluding kmod-nvidia-410.66-1.el7_5.elrepo.x86_64 >> [nvidia]: excluding kmod-nvidia-410.73-1.el7_5.elrepo.x86_64 >> [nvidia]: excluding kmod-nvidia-410.73-2.el7_6.elrepo.x86_64 >> [nvidia]: excluding
2020 Jan 06
2
[PATCH v2 0/3] drm/nouveau: Support NVIDIA format modifiers
On 1/5/20 5:30 PM, Ben Skeggs wrote: > On Tue, 17 Dec 2019 at 10:44, James Jones <jajones at nvidia.com> wrote: >> >> This series modifies the NV5x+ nouveau display backends to advertise >> appropriate format modifiers on their display planes in atomic mode >> setting blobs. >> >> Corresponding modifications to Mesa/userspace are available here:
2013 Nov 21
2
H.264 engine differences between fermi and tesla cards
On 11/19/2013 08:16 PM, Ilia Mirkin wrote: > Hello, > > I hope this is an appropriate style of request for this forum. I added > code to support video decoding on the tesla cards that have a > similar-style video decoding engine to fermi cards (i.e. G98, GT21x, > the IGP's -- the falcon-controlled decoding engines, rather than the > xtensa-controlled ones), by using