search for: reimports

Displaying 20 results from an estimated 64 matches for "reimports".

Did you mean: reimport
2023 Jul 15
1
DNS Full export and reimport
Hi all In this page of the Samba Wiki https://wiki.samba.org/index.php/Samba_AD_DC_Troubleshooting There is this note re a domain join issue. A full re-import of your DNS database via PowerShell is one way to ensure that DNS records are only in the modern locations. Can you tell me how to do this? I have found several articles but I am unsure of the exact method. Thank you very much. Travis
2019 Apr 29
2
[Spice-devel] [PATCH] Revert "drm/qxl: drop prime import/export callbacks"
On Fri, Apr 26, 2019 at 04:21:37PM +0200, Daniel Vetter wrote: > On Fri, Apr 26, 2019 at 7:33 AM Gerd Hoffmann <kraxel at redhat.com> wrote: > > > > This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8. > > > > Simliar to commit a0cecc23cfcb Revert "drm/virtio: drop prime > > import/export callbacks". We have to do the same with qxl, >
2019 Apr 29
2
[Spice-devel] [PATCH] Revert "drm/qxl: drop prime import/export callbacks"
On Fri, Apr 26, 2019 at 04:21:37PM +0200, Daniel Vetter wrote: > On Fri, Apr 26, 2019 at 7:33 AM Gerd Hoffmann <kraxel at redhat.com> wrote: > > > > This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8. > > > > Simliar to commit a0cecc23cfcb Revert "drm/virtio: drop prime > > import/export callbacks". We have to do the same with qxl, >
2007 Aug 14
2
IO error on mount for encrypted dataset
Does the ARC get flushed for a dataset when it is unmounted ? What does change when a dataset is unmounted ? The context of the problem is this: create a pool, provide the pool encryption key, create a dataset with encryption turned on, put data into that dataset I see it getting encrypted and written to disk by zio_write, zfs umount -a zfs mount -a I can read the data back - yeah!. However
2006 Sep 01
2
embed image (png) in postscript (device)
Hi, I output multiple "grid-package-based" plots to the postscript device. Because the graphics are complicated and consists of a lot of datapoints (~200'000) the files become really big. To avoid this big files and to shorten the creation, I currently print the plots to the png device and manually combine them into one multipaged pdf document. My question is, it is possible to do
2019 Apr 26
3
[PATCH] Revert "drm/qxl: drop prime import/export callbacks"
This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8. Simliar to commit a0cecc23cfcb Revert "drm/virtio: drop prime import/export callbacks". We have to do the same with qxl, for the same reasons (it breaks DRI3). Drop the WARN_ON_ONCE(). Fixes: f4c34b1e2a37d5676180901fa6ff188bcb6371f8 Signed-off-by: Gerd Hoffmann <kraxel at redhat.com> ---
2019 Apr 26
3
[PATCH] Revert "drm/qxl: drop prime import/export callbacks"
This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8. Simliar to commit a0cecc23cfcb Revert "drm/virtio: drop prime import/export callbacks". We have to do the same with qxl, for the same reasons (it breaks DRI3). Drop the WARN_ON_ONCE(). Fixes: f4c34b1e2a37d5676180901fa6ff188bcb6371f8 Signed-off-by: Gerd Hoffmann <kraxel at redhat.com> ---
2011 Feb 03
2
tapply output as a dataframe
On Mon, Apr 13, 2009 at 12:41 PM, Dan Dube <ddube-at-advisen.com> wrote: > i use tapply and by often, but i always end up banging my head against > the wall with the output. The proposed solution of Dan's problem posted on R-help was: > do.call(rbind,a) When I use this 'solution' I get 'ERROR: second argument must be a list'. So head on wall continues. My
2012 Nov 29
2
Deleting certain observations (and their imprint?)
I'm manipulating a large dataset and need to eliminate some observations based on specific identifiers. This isn't a problem in and of itself (using which.. or subset..) but an imprint of the deleted observations seem to remain, even though they have 0 observations. This is causing me problems later on. I'll use the dataset warpbreaks to illustrate, I apologize if this isn't in
2019 Apr 29
0
[Spice-devel] [PATCH] Revert "drm/qxl: drop prime import/export callbacks"
On Mon, Apr 29, 2019 at 9:54 AM Gerd Hoffmann <kraxel at redhat.com> wrote: > > On Fri, Apr 26, 2019 at 04:21:37PM +0200, Daniel Vetter wrote: > > On Fri, Apr 26, 2019 at 7:33 AM Gerd Hoffmann <kraxel at redhat.com> wrote: > > > > > > This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8. > > > > > > Simliar to commit a0cecc23cfcb
2019 Apr 29
1
[Spice-devel] [PATCH] Revert "drm/qxl: drop prime import/export callbacks"
Hi, > > More useful would be some way to signal this self-reimport capability > > to userspace somehow. See DRM_PRIME_CAP_LOCAL patch. > > Userspace is supposed to test whether import/export works for a > specific combo, not blindly assume it does and then keel over. I think > we need to fix that, not add more flags - there's lots of reasons why > a given pair of
2019 Dec 06
2
Backing up tdb files
Mandi! Rowland penny via samba In chel di` si favelave... > Then you only need to backup your smb.conf and your LDAP, the tdb files will > be recreated from ldap. AFAIK minus: a) 'smbpasswd -w', samba need to know how to access OpenLDAP. ;-) b) rights ('net rpc rights'). c) printers (per se, but in particular printer drivers), if you use it d) policy ('pdbedit
2002 Jul 25
1
OpenSSH 3.4p1's top level .cvsignore file
anyone who can help me understand- one of my developers reimported openssh snapshot from 0722 into cvs. he sent me the following email about an error he believes is in the snapshot i'm not a cvs-knowledgeable person, so i'm not all that clear on this, but should *.in files be in .cvsignore? i tried checking a later snapshot, but 0722 seems to be the latest one. .cvsignore is dated jun
2004 Mar 25
1
Convert Eror
I upgraded from Redhat 8 using LDAP 2.0.22 and Samba 2.28a to Fedora Core 1 .Reimported my LDAP Schema into the new version of Ldap but when attempting to run the following .. I recieve this error.. Any ideas? ./convertSambaAccount --input /usr/share/doc/samba-3.0.0/LDAP/txt.ldif --output new.ldif --sid S-1-5-21-387053420-3 --changetype modify Can't call method "get_value" on an
2013 Sep 19
1
AstDB Partial Replication?
Is anyone aware of a way to replicate parts of the AstDB to another Asterisk install? For example, to export all CF entries on a FreePBX based system to another system running FreePBX, I might do: asterisk -rx 'database show' | grep CF This gives me a list of data, which I can rsync to another host to reimport using 'database put'. BUT, the problem comes in when I want to sync
2008 Jul 14
3
Data Manipulations and SQL
Greetings, I am new to R and have some background knowledge about SQL. I'd like to know whether there is a way to manipulate the R datasets (or data frames) using SQL statements. For example, I have two data frames and both of them have a column called "id", then I want to join this two data frames into one. In SQL, we can just simply use the join comment. What should we do in R?
2003 Feb 07
3
RMySQL
Being relatively new to R, I want to use R in connection with a MySQL database. After installing (apparently sucessfully) the package RMySQL (R Version 1.6.2) the result is as follows: > library(RMySQL) > con <- dbConnect(MySQL(), dbname= "test") Error: couldn''t find function "dbConnect" Thank You in advance for help Peter Jörg Peter Baresel Technische
2019 Apr 26
1
[PATCH] Revert "drm/qxl: drop prime import/export callbacks"
On 4/26/19 4:21 PM, Daniel Vetter wrote: > On Fri, Apr 26, 2019 at 7:33 AM Gerd Hoffmann <kraxel at redhat.com> wrote: >> This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8. >> >> Simliar to commit a0cecc23cfcb Revert "drm/virtio: drop prime >> import/export callbacks". We have to do the same with qxl, >> for the same reasons (it breaks
2019 Apr 26
1
[PATCH] Revert "drm/qxl: drop prime import/export callbacks"
On 4/26/19 4:21 PM, Daniel Vetter wrote: > On Fri, Apr 26, 2019 at 7:33 AM Gerd Hoffmann <kraxel at redhat.com> wrote: >> This reverts commit f4c34b1e2a37d5676180901fa6ff188bcb6371f8. >> >> Simliar to commit a0cecc23cfcb Revert "drm/virtio: drop prime >> import/export callbacks". We have to do the same with qxl, >> for the same reasons (it breaks
2012 Nov 12
5
Matrix to data frame conversion
I have a matrix which I wanted to convert to a data frame. As I could not succeed and resorted to export to csv and reimport it again. Why did I fail in the attempt and how can I achieve what I wanted without this roundabouts? The original matrix: > str(comb_model0) num [1:90, 1:4] 3.5938 0.0274 0.0342 0.0135 0.0207 ... - attr(*, "dimnames")=List of 2 ..$ : chr [1:90]