Displaying 20 results from an estimated 64 matches for "reimporting".
2023 Jul 15
1
DNS Full export and reimport
Hi all
In this page of the Samba Wiki
https://wiki.samba.org/index.php/Samba_AD_DC_Troubleshooting
There is this note re a domain join issue.
A full re-import of your DNS database via PowerShell is one way to ensure
that DNS records are only in the modern locations.
Can you tell me how to do this? I have found several articles but I am
unsure of the exact method.
Thank you very much.
Travis
2019 Apr 29
2
[Spice-devel] [PATCH] Revert "drm/qxl: drop prime import/export callbacks"
On Fri, Apr 26, 2019 at 04:21:37PM +0200, Daniel Vetter wrote:
> On Fri, Apr 26, 2019 at 7:33 AM Gerd Hoffmann <kraxel at redhat.com> wrote:
> >
> > This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8.
> >
> > Simliar to commit a0cecc23cfcb Revert "drm/virtio: drop prime
> > import/export callbacks". We have to do the same with qxl,
>
2019 Apr 29
2
[Spice-devel] [PATCH] Revert "drm/qxl: drop prime import/export callbacks"
On Fri, Apr 26, 2019 at 04:21:37PM +0200, Daniel Vetter wrote:
> On Fri, Apr 26, 2019 at 7:33 AM Gerd Hoffmann <kraxel at redhat.com> wrote:
> >
> > This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8.
> >
> > Simliar to commit a0cecc23cfcb Revert "drm/virtio: drop prime
> > import/export callbacks". We have to do the same with qxl,
>
2007 Aug 14
2
IO error on mount for encrypted dataset
Does the ARC get flushed for a dataset when it is unmounted ?
What does change when a dataset is unmounted ?
The context of the problem is this:
create a pool,
provide the pool encryption key,
create a dataset with encryption turned on,
put data into that dataset
I see it getting encrypted and written to disk by zio_write,
zfs umount -a
zfs mount -a
I can read the data back - yeah!.
However
2006 Sep 01
2
embed image (png) in postscript (device)
Hi,
I output multiple "grid-package-based" plots to the postscript device.
Because the graphics are complicated and consists of a lot of datapoints
(~200'000) the files become really big. To avoid this big files and to
shorten the creation, I currently print the plots to the png device and
manually combine them into one multipaged pdf document.
My question is, it is possible to do
2019 Apr 26
3
[PATCH] Revert "drm/qxl: drop prime import/export callbacks"
This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8.
Simliar to commit a0cecc23cfcb Revert "drm/virtio: drop prime
import/export callbacks". We have to do the same with qxl,
for the same reasons (it breaks DRI3).
Drop the WARN_ON_ONCE().
Fixes: f4c34b1e2a37d5676180901fa6ff188bcb6371f8
Signed-off-by: Gerd Hoffmann <kraxel at redhat.com>
---
2019 Apr 26
3
[PATCH] Revert "drm/qxl: drop prime import/export callbacks"
This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8.
Simliar to commit a0cecc23cfcb Revert "drm/virtio: drop prime
import/export callbacks". We have to do the same with qxl,
for the same reasons (it breaks DRI3).
Drop the WARN_ON_ONCE().
Fixes: f4c34b1e2a37d5676180901fa6ff188bcb6371f8
Signed-off-by: Gerd Hoffmann <kraxel at redhat.com>
---
2011 Feb 03
2
tapply output as a dataframe
On Mon, Apr 13, 2009 at 12:41 PM, Dan Dube <ddube-at-advisen.com> wrote:
> i use tapply and by often, but i always end up banging my head against
> the wall with the output.
The proposed solution of Dan's problem posted on R-help was:
> do.call(rbind,a)
When I use this 'solution' I get 'ERROR: second argument must be a list'. So head on wall continues.
My
2012 Nov 29
2
Deleting certain observations (and their imprint?)
I'm manipulating a large dataset and need to eliminate some observations based on specific identifiers. This isn't a problem in and of itself (using which.. or subset..) but an imprint of the deleted observations seem to remain, even though they have 0 observations. This is causing me problems later on. I'll use the dataset warpbreaks to illustrate, I apologize if this isn't in
2019 Apr 29
0
[Spice-devel] [PATCH] Revert "drm/qxl: drop prime import/export callbacks"
On Mon, Apr 29, 2019 at 9:54 AM Gerd Hoffmann <kraxel at redhat.com> wrote:
>
> On Fri, Apr 26, 2019 at 04:21:37PM +0200, Daniel Vetter wrote:
> > On Fri, Apr 26, 2019 at 7:33 AM Gerd Hoffmann <kraxel at redhat.com> wrote:
> > >
> > > This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8.
> > >
> > > Simliar to commit a0cecc23cfcb
2019 Apr 29
1
[Spice-devel] [PATCH] Revert "drm/qxl: drop prime import/export callbacks"
Hi,
> > More useful would be some way to signal this self-reimport capability
> > to userspace somehow. See DRM_PRIME_CAP_LOCAL patch.
>
> Userspace is supposed to test whether import/export works for a
> specific combo, not blindly assume it does and then keel over. I think
> we need to fix that, not add more flags - there's lots of reasons why
> a given pair of
2019 Dec 06
2
Backing up tdb files
Mandi! Rowland penny via samba
In chel di` si favelave...
> Then you only need to backup your smb.conf and your LDAP, the tdb files will
> be recreated from ldap.
AFAIK minus:
a) 'smbpasswd -w', samba need to know how to access OpenLDAP. ;-)
b) rights ('net rpc rights').
c) printers (per se, but in particular printer drivers), if you use it
d) policy ('pdbedit
2002 Jul 25
1
OpenSSH 3.4p1's top level .cvsignore file
anyone who can help me understand-
one of my developers reimported openssh snapshot from 0722
into cvs. he sent me the following email about an error he believes
is in the snapshot
i'm not a cvs-knowledgeable person, so i'm not all that
clear on this, but should *.in files be in .cvsignore?
i tried checking a later snapshot, but 0722 seems to be the latest one.
.cvsignore is dated jun
2004 Mar 25
1
Convert Eror
I upgraded from Redhat 8 using LDAP 2.0.22 and Samba 2.28a to Fedora Core 1
.Reimported my LDAP Schema into the new version of Ldap but when attempting
to run the following .. I recieve this error..
Any ideas?
./convertSambaAccount --input
/usr/share/doc/samba-3.0.0/LDAP/txt.ldif --output new.ldif --sid
S-1-5-21-387053420-3 --changetype modify
Can't call method "get_value" on an
2013 Sep 19
1
AstDB Partial Replication?
Is anyone aware of a way to replicate parts of the AstDB to another Asterisk install?
For example, to export all CF entries on a FreePBX based system to another system running FreePBX, I might do:
asterisk -rx 'database show' | grep CF
This gives me a list of data, which I can rsync to another host to reimport using 'database put'. BUT, the problem comes in when I want to sync
2008 Jul 14
3
Data Manipulations and SQL
Greetings,
I am new to R and have some background knowledge about SQL. I'd like to
know whether there is a way to manipulate the R datasets (or data
frames) using SQL statements. For example, I have two data frames and
both of them have a column called "id", then I want to join this two
data frames into one. In SQL, we can just simply use the join comment.
What should we do in R?
2003 Feb 07
3
RMySQL
Being relatively new to R, I want to use R in connection with a MySQL database.
After installing (apparently sucessfully) the package RMySQL (R Version 1.6.2)
the result is as follows:
> library(RMySQL)
> con <- dbConnect(MySQL(), dbname= "test")
Error: couldn''t find function "dbConnect"
Thank You in advance for help
Peter
Jörg Peter Baresel
Technische
2019 Apr 26
1
[PATCH] Revert "drm/qxl: drop prime import/export callbacks"
On 4/26/19 4:21 PM, Daniel Vetter wrote:
> On Fri, Apr 26, 2019 at 7:33 AM Gerd Hoffmann <kraxel at redhat.com> wrote:
>> This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8.
>>
>> Simliar to commit a0cecc23cfcb Revert "drm/virtio: drop prime
>> import/export callbacks". We have to do the same with qxl,
>> for the same reasons (it breaks
2019 Apr 26
1
[PATCH] Revert "drm/qxl: drop prime import/export callbacks"
On 4/26/19 4:21 PM, Daniel Vetter wrote:
> On Fri, Apr 26, 2019 at 7:33 AM Gerd Hoffmann <kraxel at redhat.com> wrote:
>> This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8.
>>
>> Simliar to commit a0cecc23cfcb Revert "drm/virtio: drop prime
>> import/export callbacks". We have to do the same with qxl,
>> for the same reasons (it breaks
2012 Nov 12
5
Matrix to data frame conversion
I have a matrix which I wanted to convert to a data frame. As I could not
succeed and resorted to export to csv and reimport it again. Why did I fail
in the attempt and how can I achieve what I wanted without this
roundabouts?
The original matrix:
> str(comb_model0)
num [1:90, 1:4] 3.5938 0.0274 0.0342 0.0135 0.0207 ...
- attr(*, "dimnames")=List of 2
..$ : chr [1:90]