search for: waiters

Displaying 20 results from an estimated 270 matches for "waiters".

Did you mean: waiter
2020 Mar 23
2
Samba still DNS Exit Code 23
Hello Samba Group Today installing the Samba Version 4.11.7-Debian pkg from Louis, Thanks Louis !! The Server will running but watching on the service i have DNS Problmes. Please can give me any update here : what are runing wrong. -- samba-ad-dc.service - Samba AD Daemon Loaded: loaded (/lib/systemd/system/samba-ad-dc.service; enabled; vendor preset: enabled) Active:
2020 Mar 23
0
Samba still DNS Exit Code 23
Your welkom Can you run this : https://raw.githubusercontent.com/thctlo/samba4/master/samba-collect-debug-info.sh Anonimize where needed. That should tell use enough to see whats going on. Greetz, Louis > -----Oorspronkelijk bericht----- > Van: samba [mailto:samba-bounces at lists.samba.org] Namens > mauri via samba > Verzonden: maandag 23 maart 2020 14:13 > Aan:
2020 Oct 18
2
samba start issues after classic upgrade
Hello, Just completed a classic upgrade on a new server. NT4 style PDC was running 4.6.5, new server running 4.13.0 (Debian Buster compiled from source) Classic Upgrade process seemed to go fine with one caveat, not sure if it means anything - the wiki shows the upgrade process ending with: =========================== Commiting 'add users to groups' transaction to disk Setting password
2020 Mar 23
2
Samba still DNS Exit Code 23
Meny thanks for our fast answer ! root at AD:/source# cat /tmp/samba-debug-info.txt Collected config --- 2020-03-23-15:57 ----------- Hostname: AD DNS Domain: Caloro.m FQDN: AD.Caloro.m ipaddress: 192.168.1.7 ----------- Kerberos SRV _kerberos._tcp.Caloro.m record verified ok, sample output: Server: 192.168.1.9 Address: 192.168.1.9#53 _kerberos._tcp.caloro.m service = 0 0 88
2014 Mar 13
1
[PATCH RFC v6 09/11] pvqspinlock, x86: Add qspinlock para-virtualization support
...aolo Bonzini wrote: > Il 13/03/2014 12:21, David Vrabel ha scritto: >> On 12/03/14 18:54, Waiman Long wrote: >>> This patch adds para-virtualization support to the queue spinlock in >>> the same way as was done in the PV ticket lock code. In essence, the >>> lock waiters will spin for a specified number of times (QSPIN_THRESHOLD >>> = 2^14) and then halted itself. The queue head waiter will spins >>> 2*QSPIN_THRESHOLD times before halting itself. When it has spinned >>> QSPIN_THRESHOLD times, the queue head will assume that the lock >&...
2014 Mar 13
1
[PATCH RFC v6 09/11] pvqspinlock, x86: Add qspinlock para-virtualization support
...aolo Bonzini wrote: > Il 13/03/2014 12:21, David Vrabel ha scritto: >> On 12/03/14 18:54, Waiman Long wrote: >>> This patch adds para-virtualization support to the queue spinlock in >>> the same way as was done in the PV ticket lock code. In essence, the >>> lock waiters will spin for a specified number of times (QSPIN_THRESHOLD >>> = 2^14) and then halted itself. The queue head waiter will spins >>> 2*QSPIN_THRESHOLD times before halting itself. When it has spinned >>> QSPIN_THRESHOLD times, the queue head will assume that the lock >&...
2014 Mar 13
3
[PATCH RFC v6 09/11] pvqspinlock, x86: Add qspinlock para-virtualization support
On 12/03/14 18:54, Waiman Long wrote: > This patch adds para-virtualization support to the queue spinlock in > the same way as was done in the PV ticket lock code. In essence, the > lock waiters will spin for a specified number of times (QSPIN_THRESHOLD > = 2^14) and then halted itself. The queue head waiter will spins > 2*QSPIN_THRESHOLD times before halting itself. When it has spinned > QSPIN_THRESHOLD times, the queue head will assume that the lock > holder may be scheduled...
2014 Mar 13
3
[PATCH RFC v6 09/11] pvqspinlock, x86: Add qspinlock para-virtualization support
On 12/03/14 18:54, Waiman Long wrote: > This patch adds para-virtualization support to the queue spinlock in > the same way as was done in the PV ticket lock code. In essence, the > lock waiters will spin for a specified number of times (QSPIN_THRESHOLD > = 2^14) and then halted itself. The queue head waiter will spins > 2*QSPIN_THRESHOLD times before halting itself. When it has spinned > QSPIN_THRESHOLD times, the queue head will assume that the lock > holder may be scheduled...
2024 Sep 11
1
Upgrade 4.19 to 4.20 defunct
I upgraded one of my DCs from 4.19 to 4.20 using Michaels repository (http://www.corpit.ru/mjt/packages/samba jammy/samba-4.20/), and the DC is now no longer responding. systemctl status samba-ad-dc returns: ? samba-ad-dc.service - LSB: Samba daemons for the AD DC Loaded: loaded (/etc/init.d/samba-ad-dc; generated) Active: active (exited) since Wed 2024-09-11 02:38:32 UTC; 3h 29min
2024 Sep 11
1
Upgrade 4.19 to 4.20 defunct
From http://samba.bigbird.es/doku.php?id=samba:upgrade-sama 4.20 has a new package samba-ad-dc (split from other binaries in previous versions) that provides DC functionality. Before 4.20, there is a meta-package with the same name that provides compatibility before the upgrade. It is recommended to install this meta package samba-ad-dc before upgrading to 4.20, so the upgrade has all its
2020 Nov 04
2
Active - Deactivating
No installed Debian 10.6, running local Bind9 Server, and krb5-kdc.service Running fine. Installing the Louse package ?samba 2:4.13.1+dfsg-0.1buster1? root at AD:/etc/init.d# systemctl status samba-ad-dc ? samba-ad-dc.service - Samba AD Daemon Loaded: loaded (/lib/systemd/system/samba-ad-dc.service; enabled; vendor preset: enabled) Active: deactivating (stop-sigterm) (Result:
2018 Jan 11
0
[PATCH 1/3] gpu: host1x: Add support for DMA fences
From: Mikko Perttunen <mperttunen at nvidia.com> Add an implementation of DMA fences backed by Host1x syncpoints, an interface to specify a prefence for job submissions. Before submission, prefences containing only Host1x syncpoints are waited by pushing wait commands to CDMA, whereas other fences are CPU-waited. Signed-off-by: Mikko Perttunen <mperttunen at nvidia.com>
2018 Jan 11
6
[PATCH 0/3] drm/tegra: Add support for fence FDs
From: Thierry Reding <treding at nvidia.com> This set of patches adds support for fences to Tegra DRM and complements the fence FD support for Nouveau. Technically this isn't necessary for a fence-based synchronization loop with Nouveau because the KMS core takes care of all that, but engines behind host1x can use the IOCTL extensions provided here to emit fence FDs that in turn can be
2014 Mar 13
0
[PATCH RFC v6 09/11] pvqspinlock, x86: Add qspinlock para-virtualization support
Il 13/03/2014 12:21, David Vrabel ha scritto: > On 12/03/14 18:54, Waiman Long wrote: >> This patch adds para-virtualization support to the queue spinlock in >> the same way as was done in the PV ticket lock code. In essence, the >> lock waiters will spin for a specified number of times (QSPIN_THRESHOLD >> = 2^14) and then halted itself. The queue head waiter will spins >> 2*QSPIN_THRESHOLD times before halting itself. When it has spinned >> QSPIN_THRESHOLD times, the queue head will assume that the lock >> holder m...
2010 Jul 21
2
Issues reshaping data
Hello All, I'm having some trouble reshaping my data from wide to long format. I have tried using both the reshape function and package. Although I haven't worked much with the reshape function, I have found the reshape package useful and intuitive for reshaping data from long to wide format. However, going the other way has me stumped with this type of data. My data is set up, roughly,
2014 Mar 14
4
[PATCH v6 05/11] pvqspinlock, x86: Allow unfair spinlock in a PV guest
On Thu, Mar 13, 2014 at 04:05:19PM -0400, Waiman Long wrote: > On 03/13/2014 11:15 AM, Peter Zijlstra wrote: > >On Wed, Mar 12, 2014 at 02:54:52PM -0400, Waiman Long wrote: > >>+static inline void arch_spin_lock(struct qspinlock *lock) > >>+{ > >>+ if (static_key_false(&paravirt_unfairlocks_enabled)) > >>+ queue_spin_lock_unfair(lock); >
2014 Mar 14
4
[PATCH v6 05/11] pvqspinlock, x86: Allow unfair spinlock in a PV guest
On Thu, Mar 13, 2014 at 04:05:19PM -0400, Waiman Long wrote: > On 03/13/2014 11:15 AM, Peter Zijlstra wrote: > >On Wed, Mar 12, 2014 at 02:54:52PM -0400, Waiman Long wrote: > >>+static inline void arch_spin_lock(struct qspinlock *lock) > >>+{ > >>+ if (static_key_false(&paravirt_unfairlocks_enabled)) > >>+ queue_spin_lock_unfair(lock); >
2014 May 14
2
[PATCH v10 03/19] qspinlock: Add pending bit
2014-05-13 15:47-0400, Waiman Long: > On 05/12/2014 11:22 AM, Radim Kr?m?? wrote: > >I think there is an unwanted scenario on virtual machines: > >1) VCPU sets the pending bit and start spinning. > >2) Pending VCPU gets descheduled. > > - we have PLE and lock holder isn't running [1] > > - the hypervisor randomly preempts us > >3) Lock holder
2014 May 14
2
[PATCH v10 03/19] qspinlock: Add pending bit
2014-05-13 15:47-0400, Waiman Long: > On 05/12/2014 11:22 AM, Radim Kr?m?? wrote: > >I think there is an unwanted scenario on virtual machines: > >1) VCPU sets the pending bit and start spinning. > >2) Pending VCPU gets descheduled. > > - we have PLE and lock holder isn't running [1] > > - the hypervisor randomly preempts us > >3) Lock holder
2014 Jul 01
2
[RFC PATCH v2] Implement Batched (group) ticket lock
On 07/01/2014 01:35 PM, Peter Zijlstra wrote: > On Sat, Jun 28, 2014 at 02:47:04PM +0530, Raghavendra K T wrote: >> In virtualized environment there are mainly three problems >> related to spinlocks that affects performance. >> 1. LHP (lock holder preemption) >> 2. Lock Waiter Preemption (LWP) >> 3. Starvation/fairness >> >> Though Ticketlocks solve