Displaying 20 results from an estimated 130 matches for "hpd".
Did you mean:
had
2019 Sep 13
1
[PATCH v2 23/27] drm/amdgpu: Iterate through DRM connectors correctly
...tor_iter(connector, &iter)
> amdgpu_connector_hotplug(connector);
> + drm_connector_list_iter_end(&iter);
> mutex_unlock(&mode_config->mutex);
> /* Just fire off a uevent and let userspace tell us what to do */
> drm_helper_hpd_irq_event(dev);
> diff --git a/drivers/gpu/drm/amd/amdgpu/dce_v10_0.c b/drivers/gpu/drm/amd/amdgpu/dce_v10_0.c
> index 645550e7caf5..be82871ac3bd 100644
> --- a/drivers/gpu/drm/amd/amdgpu/dce_v10_0.c
> +++ b/drivers/gpu/drm/amd/amdgpu/dce_v10_0.c
> @@ -330,9 +330,11 @@ static void dc...
2019 Sep 03
0
[PATCH v2 23/27] drm/amdgpu: Iterate through DRM connectors correctly
...nnector_list_iter_begin(dev, &iter);
+ drm_for_each_connector_iter(connector, &iter)
amdgpu_connector_hotplug(connector);
+ drm_connector_list_iter_end(&iter);
mutex_unlock(&mode_config->mutex);
/* Just fire off a uevent and let userspace tell us what to do */
drm_helper_hpd_irq_event(dev);
diff --git a/drivers/gpu/drm/amd/amdgpu/dce_v10_0.c b/drivers/gpu/drm/amd/amdgpu/dce_v10_0.c
index 645550e7caf5..be82871ac3bd 100644
--- a/drivers/gpu/drm/amd/amdgpu/dce_v10_0.c
+++ b/drivers/gpu/drm/amd/amdgpu/dce_v10_0.c
@@ -330,9 +330,11 @@ static void dce_v10_0_hpd_init(struct a...
2011 Jun 19
1
Multivariate HPD credible volume -- is it computable in R?
Hi all,
I'm new to the list and am hoping to get some advice. I have a set of
multivariate data and would like to find the densest part of the data cloud
containing 95% of the data, like a 95% HPD credible volume. Is there any R
code available to compute that?
Thank you very much! Your help and patience are much appreciated.
G.S.
[[alternative HTML version deleted]]
2013 May 08
1
How to calculate Hightest Posterior Density (HPD) of coeficients in a simple regression (lm) in R?
Hi!
I am trying to calculate HPD for the coeficients of regression models
fitted with lm or lmrob in R, pretty much in the same way that can be
accomplished by the association of mcmcsamp and HPDinterval functions for
multilevel models fitted with lmer. Can anyone point me in the right
direction on which packages/how to implement...
2013 May 26
2
Commit "drm: run the hpd irq event code directly" causes stutter from repeated EDID retrievals
...ently
> causes repeated EDID retrievals and failed checksums. I did a git bisect which
> lead me to
> ***
> commit 69787f7da6b2adc4054357a661aaa1701a9ca76f
> Author: Daniel Vetter <daniel.vetter at ffwll.ch>
> Date: Tue Oct 23 18:23:34 2012 +0000
>
> drm: run the hpd irq event code directly
>
> All drivers already have a work item to run the hpd code, so we don't
> need to launch a new one in the helper code. Dave Airlie mentioned
> that the cancel+re-queue might paper over DP related hpd ping-pongs,
> hence why this is split...
2019 Sep 13
0
[PATCH v2 24/27] drm/amdgpu/dm: Resume short HPD IRQs before resuming MST topology
On Tue, Sep 3, 2019 at 4:49 PM Lyude Paul <lyude at redhat.com> wrote:
>
> Since we're going to be reprobing the entire topology state on resume
> now using sideband transactions, we need to ensure that we actually have
> short HPD irqs enabled before calling drm_dp_mst_topology_mgr_resume().
> So, do that.
>
> Cc: Juston Li <juston.li at intel.com>
> Cc: Imre Deak <imre.deak at intel.com>
> Cc: Ville Syrjälä <ville.syrjala at linux.intel.com>
> Cc: Harry Wentland <hwentlan at amd.com>...
2019 Sep 03
0
[PATCH v2 24/27] drm/amdgpu/dm: Resume short HPD IRQs before resuming MST topology
Since we're going to be reprobing the entire topology state on resume
now using sideband transactions, we need to ensure that we actually have
short HPD irqs enabled before calling drm_dp_mst_topology_mgr_resume().
So, do that.
Cc: Juston Li <juston.li at intel.com>
Cc: Imre Deak <imre.deak at intel.com>
Cc: Ville Syrjälä <ville.syrjala at linux.intel.com>
Cc: Harry Wentland <hwentlan at amd.com>
Cc: Daniel Vetter <danie...
2008 Sep 19
1
readRegistry function (PR#12937)
Full_Name: Zivan Karaman
Version: 2.7.2
OS: Windows XP
Submission from: (NULL) (195.6.68.214)
I'm puzzled by the readRegistry function.
Shouldn't the "hive" argument be something like
c("HLM", "HCR", "HCU", "HU", "HCC", "HPD") rather than
c("HLM", "HCR", "HCU", "HU", "HCC, HPD").
For example,
> readRegistry("Test", "HCC, HPD")
Error in readRegistry("Test", "HCC, HPD") : invalid 'hive' value
Please note also tha...
2011 Aug 23
1
pMCMC and HPD in MCMCglmm
...st.mean l-95% CIu-95% CI eff.samp pMCMC
(Intercept) -3.6332 -5.6136 -1.7719 3045 <2e-04 ***
as.factor(sex)M -2.9959 -6.0690 0.1969 2628 0.0455 *
---
Signif. codes: 0 ?***?0.001 ?**? 0.01 ?*? 0.05 ?.? 0.1 ? ? 1
As you can see, pMCMC for gender is just less than 5%, but the credible
interval (HPD) is wide and includes the 0 value.
How can I interpret these different results?
Thank you in advance
Massimo
-----------------------
Massimo Fenati
DVM
Padova - Italy
2018 Aug 06
2
[PATCH v3 3/8] drm/fb_helper: Introduce hotplug_suspend/resume()
...tion in the review of one of my other patches that we should avoid
> disabling polling during runtime suspend, and you're definitely right. I feel
> a bit silly for not remembering that since I was the one who made it so that
> i915 does polling in runtime suspend for chips without RPM HPD detection in
> the first place because it was causing people's displays not to come up on
> vlv...
> Anyway: I think if we just leave output polling enabled during runtime suspend
> that might actually fix all of the fb_helper locking issues since we won't
> need to wait on a...
2018 Aug 01
0
[PATCH v4 8/8] drm/nouveau: Call pm_runtime_get_noresume() from hpd handlers
...= connector->name;
struct nouveau_encoder *nv_encoder;
+ /* Resuming the device here isn't possible; but the suspend PM ops
+ * will wait for us to finish our work before disabling us so this
+ * should be enough
+ */
+ pm_runtime_get_noresume(drm->dev->dev);
nv_connector->hpd_task = current;
if (rep->mask & NVIF_NOTIFY_CONN_V0_IRQ) {
@@ -1171,6 +1176,9 @@ nouveau_connector_hotplug(struct nvif_notify *notify)
}
nv_connector->hpd_task = NULL;
+
+ pm_runtime_mark_last_busy(drm->dev->dev);
+ pm_runtime_put_autosuspend(drm->dev->dev);
return...
2018 Aug 07
0
[PATCH v5 13/13] drm/nouveau: Call pm_runtime_get_noresume() from hpd handlers
...= connector->name;
struct nouveau_encoder *nv_encoder;
+ /* Resuming the device here isn't possible; but the suspend PM ops
+ * will wait for us to finish our work before disabling us so this
+ * should be enough
+ */
+ pm_runtime_get_noresume(drm->dev->dev);
nv_connector->hpd_task = current;
if (rep->mask & NVIF_NOTIFY_CONN_V0_IRQ) {
@@ -1171,6 +1176,9 @@ nouveau_connector_hotplug(struct nvif_notify *notify)
}
nv_connector->hpd_task = NULL;
+
+ pm_runtime_mark_last_busy(drm->dev->dev);
+ pm_runtime_put_autosuspend(drm->dev->dev);
return...
2023 Aug 14
1
2b5d1c29f6c4 ("drm/nouveau/disp: PIOR DP uses GPIO for HPD, not PMGR AUX interrupts")
On Mon, 14 Aug 2023 16:51:08 +0200,
Karol Herbst wrote:
>
> I've sent a patch out to address this memory corruption
> https://patchwork.freedesktop.org/patch/552642/
>
> It might or might not fix regressions from the original I2C fix, so
> please test and report if there are remaining issues.
Thanks! I'll build a test kernel and ask the reporter for testing
with it.
2019 Jul 18
0
[PATCH 21/26] drm/nouveau: Don't grab runtime PM refs for HPD IRQs
...make sure
that nouveau doesn't bother grabbing a runtime PM reference to do so,
since otherwise we'll start deadlocking runtime PM again.
Note that we weren't able to do this before, because of the DP MST
helpers processing UP requests from topologies in the same context as
drm_dp_mst_hpd_irq() which would have caused us to open ourselves up to
receiving hotplug events and deadlocking with runtime suspend/resume.
Now that those requests are handled asynchronously, this change should
be completely safe.
Cc: Juston Li <juston.li at intel.com>
Cc: Imre Deak <imre.deak at inte...
2019 Sep 03
0
[PATCH v2 22/27] drm/nouveau: Don't grab runtime PM refs for HPD IRQs
...make sure
that nouveau doesn't bother grabbing a runtime PM reference to do so,
since otherwise we'll start deadlocking runtime PM again.
Note that we weren't able to do this before, because of the DP MST
helpers processing UP requests from topologies in the same context as
drm_dp_mst_hpd_irq() which would have caused us to open ourselves up to
receiving hotplug events and deadlocking with runtime suspend/resume.
Now that those requests are handled asynchronously, this change should
be completely safe.
Cc: Juston Li <juston.li at intel.com>
Cc: Imre Deak <imre.deak at inte...
2019 Oct 22
0
[PATCH v5 09/14] drm/nouveau: Don't grab runtime PM refs for HPD IRQs
...make sure
that nouveau doesn't bother grabbing a runtime PM reference to do so,
since otherwise we'll start deadlocking runtime PM again.
Note that we weren't able to do this before, because of the DP MST
helpers processing UP requests from topologies in the same context as
drm_dp_mst_hpd_irq() which would have caused us to open ourselves up to
receiving hotplug events and deadlocking with runtime suspend/resume.
Now that those requests are handled asynchronously, this change should
be completely safe.
Cc: Juston Li <juston.li at intel.com>
Cc: Imre Deak <imre.deak at inte...
2018 Aug 16
0
[PATCH] drm/nouveau: Prevent handling ACPI HPD events too early
On most systems with ACPI hotplugging support, it seems that we always
receive a hotplug event once we re-enable EC interrupts even if the GPU
hasn't even been resumed yet.
This can cause problems since even though we schedule hpd_work to handle
connector reprobing for us, hpd_work synchronizes on
pm_runtime_get_sync() to wait until the device is ready to perform
reprobing. Since runtime suspend/resume callbacks are disabled before
the PM core calls ->suspend(), any calls to pm_runtime_get_sync() during
this period will g...
2019 Dec 10
0
[PATCH AUTOSEL 5.4 146/350] drm/nouveau: Don't grab runtime PM refs for HPD IRQs
...make sure
that nouveau doesn't bother grabbing a runtime PM reference to do so,
since otherwise we'll start deadlocking runtime PM again.
Note that we weren't able to do this before, because of the DP MST
helpers processing UP requests from topologies in the same context as
drm_dp_mst_hpd_irq() which would have caused us to open ourselves up to
receiving hotplug events and deadlocking with runtime suspend/resume.
Now that those requests are handled asynchronously, this change should
be completely safe.
Cc: Juston Li <juston.li at intel.com>
Cc: Imre Deak <imre.deak at inte...
2018 Aug 02
1
[PATCH v4 8/8] drm/nouveau: Call pm_runtime_get_noresume() from hpd handlers
...veau_encoder *nv_encoder;
>
> + /* Resuming the device here isn't possible; but the suspend PM ops
> + * will wait for us to finish our work before disabling us so this
> + * should be enough
> + */
> + pm_runtime_get_noresume(drm->dev->dev);
> nv_connector->hpd_task = current;
>
> if (rep->mask & NVIF_NOTIFY_CONN_V0_IRQ) {
> @@ -1171,6 +1176,9 @@ nouveau_connector_hotplug(struct nvif_notify *notify)
> }
>
> nv_connector->hpd_task = NULL;
> +
> + pm_runtime_mark_last_busy(drm->dev->dev);
> + pm_runtime_p...
2023 Aug 14
2
[PATCH] drm/nouveau/disp: fix use-after-free in error handling of nouveau_connector_create
We can't simply free the connector after calling drm_connector_init on it.
We need to clean up the drm side first.
It might not fix all regressions from 2b5d1c29f6c4 ("drm/nouveau/disp:
PIOR DP uses GPIO for HPD, not PMGR AUX interrupts"), but at least it
fixes a memory corruption in error handling related to that commit.
Link: https://lore.kernel.org/lkml/20230806213107.GFZNARG6moWpFuSJ9W at fat_crate.local/
Fixes: 95983aea8003 ("drm/nouveau/disp: add connector class")
Signed-off-by: Karol...