similar to: Scripted video decoding firmware extraction

Displaying 20 results from an estimated 20000 matches similar to: "Scripted video decoding firmware extraction"

2019 Aug 14
1
Video Hardware Decoding: Jittery Rectangles on Nvidia GT218 NVA8 VP4.
Hi Ilia, A fortnight ago, you wrote: > > The video plays, CPU load is less (my aim), but there's ‘tearing’ of > > the picture as if small rectangles that are updates are appearing in > > the wrong location, off by a little. If I step through the frames > > with mpv's ‘.’ and ‘,’ then I've found a pattern: one frame's > > picture is good, followed by
2019 Jul 29
1
Video Hardware Decoding: Jittery Rectangles on Nvidia GT218 NVA8 VP4.
Às 14:46 de 27/07/2019, Ilia Mirkin escreveu: > On Sat, Jul 27, 2019 at 7:37 AM Ralph Corderoy <ralph at inputplus.co.uk> wrote: >> The video plays, CPU load is less (my aim), but there's ‘tearing’ of the >> picture as if small rectangles that are updates are appearing in the >> wrong location, off by a little. If I step through the frames with >> mpv's
2019 Jul 27
0
Video Hardware Decoding: Jittery Rectangles on Nvidia GT218 NVA8 VP4.
On Sat, Jul 27, 2019 at 7:37 AM Ralph Corderoy <ralph at inputplus.co.uk> wrote: > The video plays, CPU load is less (my aim), but there's ‘tearing’ of the > picture as if small rectangles that are updates are appearing in the > wrong location, off by a little. If I step through the frames with > mpv's ‘.’ and ‘,’ then I've found a pattern: one frame's picture is
2019 Jul 27
3
Video Hardware Decoding: Jittery Rectangles on Nvidia GT218 NVA8 VP4.
Hi, I'm having trouble with hardware decoding of video with mpv(1) having switched from Nvidia's proprietary driver to nouveau due to Arch Linux removing its nvidia-340xx package. The machine is an Acer Revo. $ lspci -nn | grep VGA 01:00.0 VGA compatible controller [0300]: NVIDIA Corporation GT218 [ION] [10de:0a64] (rev a2) I think this is codename NVA8 with the
2016 Mar 01
0
GSOC 2016 Aspirant for Kepler Accelerated Video Encoding;
First, make sure there's a mentor available -- I doubt I'll have time to do it, but perhaps someone else will be able to. I would recommend writing a minimal application that e.g. encodes a fixed group of frames. You should then trace this app under the blob with both mmiotrace to figure out how to configure the underlying engine (probably just firmware to upload), as well as get a mmt
2018 Oct 19
2
8K HEVC decoding with NVIDIA GeForce GTX 1060
Hi all, I have developed an Android VR player, which supports 4K video playback so far. I want to include the support for 8K video. The current Snapdragon 845 mobile platform supports up to 4K video playback, which lacks the support for my requirement. So I choose Android-x86 as my platform for development, which is Android OS runs on Intel x86 architecture. I've tried to play 8K(7680*3840)
2013 Dec 12
1
Is binary firmware still necessary for GTX660 card (NVE0 family) in order to use DRM and/or VDPAU video acceleration?
Hi Martin, if you refer to my kernel version. 3.10.17 is the latest, stable version in the official gentoo repository for the amd64 architeture. See here [4] https://packages.gentoo.org/package/sys-kernel/gentoo-sources?full_cat As long as I do not miss any features, I stay with the stable version. Hence, if I understand you correctly, there are kernel version (newer than some unknown point in
2013 Dec 12
0
Is binary firmware still necessary for GTX660 card (NVE0 family) in order to use DRM and/or VDPAU video acceleration?
On Thu, Dec 12, 2013 at 2:32 AM, Matthias Nagel <matthias.h.nagel at gmail.com> wrote: > Hello, > > I run the Gentoo Linux distribution and use a self-compiled Linux 3.10.17 > kernel. According to > > [1] http://nouveau.freedesktop.org/wiki/InstallDRM/ > [2] http://nouveau.freedesktop.org/wiki/VideoAcceleration/ > > I need the original firmware from the binary
2013 Dec 12
0
Is binary firmware still necessary for GTX660 card (NVE0 family) in order to use DRM and/or VDPAU video acceleration?
On 12/12/2013 08:32, Matthias Nagel wrote: > Hello, > > I run the Gentoo Linux distribution and use a self-compiled Linux > 3.10.17 kernel. According to > > [1] http://nouveau.freedesktop.org/wiki/InstallDRM/ > [2] http://nouveau.freedesktop.org/wiki/VideoAcceleration/ > > I need the original firmware from the binary driver in order to > sucessully use DRM and to
2007 Nov 11
1
nv50 microcode/shader format
Hello there, With the use of CUDA and ptx I managed to make a disassembler for at least one of the nv50 microcode/shader formats. This might come in handy for some people, hence I'm posting it here: http://www.cs.rug.nl/~wladimir/decuda/ I'm not yet sure if pixel/vertex shaders use the same instruction format as the compute shaders used by CUDA, but you'd think so, as nv50 is
2013 Dec 15
1
Is binary firmware still necessary for GTX660 card (NVE0 family) in order to use DRM and/or VDPAU video acceleration?
Hi, just a note for everybody who stumble across this thread in the future in order to bring this thread to a proper end. If I run the kernel with nouveau.debug="PVP=debug,PBSP=debug,PPPP=debug" I see messages about the firmware files gettings loaded. If the firmware files are missing, I get some warnings/errors. So far everything works as expected. But if I invoke
2013 Dec 12
4
Is binary firmware still necessary for GTX660 card (NVE0 family) in order to use DRM and/or VDPAU video acceleration?
Hello, I run the Gentoo Linux distribution and use a self-compiled Linux 3.10.17 kernel. According to [1] http://nouveau.freedesktop.org/wiki/InstallDRM/ [2] http://nouveau.freedesktop.org/wiki/VideoAcceleration/ I need the original firmware from the binary driver in order to sucessully use DRM and to use VDPAU video acceleration. I used the python script from [2] and I had a look at the ebuild
2013 Oct 17
0
MmioTrace: Using the Instruction Decoder, etc.
Hi, Good to know that! Yes, it should be faster than page faulting, although I haven't done the benchmarking yet. And yes, it is not needed to disable all but one CPU. In my current implementation, I use an ordered workqueue to send the data to the mmapped output buffer (where they will be read from from the user space) and that ensures the order of events is kept. May be less than ideal but
2016 Mar 01
4
GSOC 2016 Aspirant for Kepler Accelerated Video Encoding;
Hi, First of all, congratulations to the X.org team for selection into Google Summer of Code 2016. I would like to participate to Google Summer of Code 2016 by contributing to Kepler Accelerated Video Encoding for Nouveau. I request you to kindly guide me as to how to get started with this project. Regards, Saket Sinha
2013 Oct 19
3
MmioTrace: Using the Instruction Decoder, etc.
On Fri, 18 Oct 2013 00:11:15 +0400 Eugene Shatokhin <euspectre at gmail.com> wrote: > Hi, > > Good to know that! > > Yes, it should be faster than page faulting, although I haven't done the > benchmarking yet. And yes, it is not needed to disable all but one CPU. In > my current implementation, I use an ordered workqueue to send the data to > the mmapped output
2013 Oct 19
0
MmioTrace: Using the Instruction Decoder, etc.
Hi, > Ah, you are not using the ftrace framework nor relayfs? Mmiotrace used to be relayfs at one point and then converted to ftrace. Yes, I considered these when I started working on KernelStrider but finally borrowed ideas from Perf and implemented them. A mmapped ring buffer does its job well and has a higher throughput than Ftrace in my case. > Are you saying that you intercept
2013 Dec 07
0
H.264 engine differences between fermi and tesla cards
On Sat, 30 Nov 2013 12:54:45 -0800 Ilia Mirkin <imirkin at alum.mit.edu> wrote: > On Thu, Nov 21, 2013 at 5:22 PM, Ilia Mirkin <imirkin at alum.mit.edu> > wrote: > > On Thu, Nov 21, 2013 at 5:07 PM, Benjamin Morris > > <bmorris at nvidia.com> wrote: > >> On 11/19/2013 08:16 PM, Ilia Mirkin wrote: > >>> Hello, > >>> >
2018 Oct 19
0
8K HEVC decoding with NVIDIA GeForce GTX 1060
On Fri, Oct 19, 2018 at 8:53 AM 孫偉程 <dmvictor at gmail.com> wrote: > I can imagine there is a long way to go to fulfill my requirements. But I know a driver for my NVIDIA card is a must for the hardware decoder to work. Can somebody shed some lights that if I want to play an 8K HEVC encoded video on Android-x86 with NVIDIA's decoder enabled, what should I do in this project? And how
2014 Nov 24
1
idea of extracting microcode
Hey, Probably mostly for Ben, but I had this idea that it might be possible to write a userspace wrapper for the binary driver that run as root could be used to do the card init for the current or even a fake GPU then produce traces to pull the firmwares out. Not sure how insane it is, vs having to mmiotrace, but I just thought I'd see if anyone had considered it. Dave.
2013 Oct 17
2
MmioTrace: Using the Instruction Decoder, etc.
On Mon, 14 Oct 2013 22:45:09 +0400 Eugene Shatokhin <euspectre at gmail.com> wrote: > Hi, > > There is an interesting TODO item on MmioTraceDeveloper page: > "kprobes has a generic instruction decoding facility, use that instead of > homebrewn (or KVM), and use emulation instead of page faulting" > > Actually, I have done something similar in one of my