similar to: How to Dedup a Spatial Points Data Set

Displaying 20 results from an estimated 4000 matches similar to: "How to Dedup a Spatial Points Data Set"

2013 Mar 21
3
spatstat error
Good day Im working with some coordinates, and want to create a PPP object, I found that error: > Datos=read.table("puntos_texto.txt",dec=".",sep="\t",header=T) > summary(Datos) id y x Min. : 1.0 Min. :1013581 Min. :1177842 1st Qu.: 821.2 1st Qu.:1014442 1st Qu.:1179658 Median
2006 Jan 25
1
How to use rfm.test ? (Package MarkedPointProcess)
I would like to compute the MC test (rfm.test) available in the package MarkedPointProcess (for the data BITOEK for example) in order to test the dependence between the marks and their locations. Why the syntax of rfm.test is false here? I have the message : ****************************** ML WARNING! Forbidden values! -- if there are too many warnings try narrower lower and upper bounds for the
2013 Jan 29
1
points rejected as lying outside the specified window
Hello, I am using the following code to create ppp files from csv data and map shape files, but I am getting some errors which I have been unable to fix by searching them online: library(spatstat) library(maps) library(maptools) NYC2<-readShapePoly("nybb.shp") # this is a map of the NYC boroughs without waterways and no census tract divisions (but it does include lines separating
2011 Sep 23
1
spatstat => owin + image
Dear Community I am at my wits end and seek advice. My wish is to plot coordinates (x,y in WGS84_UMTS for the ones interested) of sampling points. This I can do by the standard spatstat prodcedure via owin. I then try to add an image, which is a map/satellite photo in the background. Firstly Problem: With the current code, I can not get the edges of the image to match the boarders of the plot
2013 Apr 11
2
Encircle the closest points
Dear R-user, For the "matplot" of the following data drawn by >matplot(t, pch=1) structure(c(2372L, 2494L, 2170L, 2238L, 1918L, 715L, 1573L, 2437L, 2481L, 1881L, 440L, 2165L, 343L, 121L, 597L, 1163L, 1459L, 2093L, 1454L, 1761L, 1872L, 174L, 2040L, 552L, 1403L, 6381L, 1649L, 4136L, 4415L, 10650L, 12111L, 4397L, 9593L, 7040L, 3888L, 8340L, 11475L, 11413L, 3985L, 1077L, 7723L, 1392L,
2013 Feb 07
2
Spatstat - Kest, envelope problem
Hello, I am writing cause I have a problem when try to create confidence envelopes in spatstat. I have a point data representing firms in my study area. When I try to create confidence envelopes the error message turns up: [> p <- envelope(point, fun=Kest, nsim=99) Error in marks.ppp(Y, dfok = FALSE) : Sorry, not implemented when the marks are a data frame] Could you write me how can I
2012 Dec 07
1
Fwd: Simulation of spatial Log-Gaussian Cox process in Spatstat
Hello, I have fitted a Log-Gaussian Cox Process on my data but when I try to use "simulate.kppm" of the spatstat package I get this error: Error in rLGCP(model = model, mu = mu, param = param, ..., win = win) : The spatial domain of the pixel image ?mu? does not cover the simulation window ?win? I've used covariates as im, changed the npixel value of the spatstat options to
2009 Mar 06
1
help to use ppp in spatial analysis
Hi, i am using spatstat package for spatial data analysis and now i have a problem to create a point pattern. The points are in file "points.txt" (first column for Latitude and second column for Longitude ) and I imported them and separated each columns in two arrays x and y. If I plots x and y ( e.g plot(x,y) ) the result appears in square area without problems and the scale is
2011 Dec 02
1
Willkommen bei der "R-help" Mailingliste
Hello everbody, I am new to this mailing list and hope to find some help. I'm trying to get into the spatstat package and encountered two problems. First a graphical one: There is an example dataset called "finpines" which has several marks (http://www.oga-lab.net/RGM2/func.php?rd_id=spatstat:finpines) When I pass the given code from the website to R data(finpines)
2010 Jun 18
1
Question : Sun Storage 7000 dedup ratio per share
Dear All : Under Sun Storage 7000 system, can we see per share ratio after enable dedup function ? We would like deep to see each share dedup ratio. On Web GUI, only show dedup ratio entire storage pool. Thanks a lot, -- Rex -- This message posted from opensolaris.org
2011 Jun 29
0
SandForce SSD internal dedup
This article raises the concern that SSD controllers (in particular SandForce) do internal dedup, and in particular that this could defeat ditto-block style replication of critical metadata as done by filesystems including ZFS. http://storagemojo.com/2011/06/27/de-dup-too-much-of-good-thing/ Along with discussion of risk evaluation, it also suggests that filesystems could vary each copy in some
2011 Jan 24
0
ZFS/ARC consuming all memory on heavy reads (w/ dedup enabled)
Greetings Gentlemen, I''m currently testing a new setup for a ZFS based storage system with dedup enabled. The system is setup on OI 148, which seems quite stable w/ dedup enabled (compared to the OpenSolaris snv_136 build I used before). One issue I ran into, however, is quite baffling: With iozone set to 32 threads, ZFS''s ARC seems to consume all available memory, making
2012 Mar 16
1
Spatstat - coordinates in observation window
Dear R users, I wish to run spatial point pattern analysis (e.g. pair correlation function, mark correlation function) for which I need to create an observation window (window=owin) from which the spatial analysis is generated. The command I used to create this observation window as follows: X1<- ppp(x, y, window=owin(c(80.58,144.96),c(101.06,165.13)),
2012 May 03
1
Creating a point pattern with cartesian co-ordinates
I have the following data from an image analysis program, in which the x and y co-ordinates are locations of the centroids of shapes on a 2 dimensional plot. The Y co-ordinates were positive, but I changed them to negative as the resulting scatterplot was upside down (the image analysis program reads from the top of the image to the bottom, so it seems) I now need these as point pattern data, as I
2011 Apr 28
4
Finding where dedup''d files are
Is there an easy way to find out what datasets have dedup''d data in them. Even better would be to discover which files in a particular dataset are dedup''d. I ran # zdb -DDDD which gave output like: index 1055c9f21af63 refcnt 2 single DVA[0]=<0:1e274ec3000:2ac00:STD:1> [L0 deduplicated block] sha256 uncompressed LE contiguous unique unencrypted 1-copy size=20000L/20000P
2009 Dec 30
3
what happens to the deduptable (DDT) when you set dedup=off ???
I tried the deduplication feature but the performance of my fileserver dived from writing 50MB/s via CIFS to 4MB/s. what happens to the deduped blocks when you set dedup=off? are they written back to disk? is the deduptable deleted or is it still there? thanks -- This message posted from opensolaris.org
2007 Dec 03
1
Function to find boundary of an irregular sample?
Given a set of coordinates that form an irregular sampling area, is there an R function to determine boundary points (coordinates defining the limits of the area), either with or without user interaction ? # for example, given the following irregular sampling area, how could I define the boundary polygon? library(SemiPar) data(scallop) plot(scallop$longitude, scallop$latitude) Thanks. --Dale
2009 Jun 16
2
dedup in dovecot?
Hi all Deduplicating data is not really a new thing, but quite efficient in mail systems where an email with an nMB attachment may be sent to multiple recipients. This might call for deduplicating data. Is there a way to do this, or is it far off? If I understand the system correctly, usually an MTA is calling dovecot on every single message, meaning the message itself won't
2010 Sep 25
4
dedup testing?
Hi all Has anyone done any testing with dedup with OI? On opensolaris there is a nifty "feature" that allows the system to hang for hours or days if attempting to delete a dataset on a deduped pool. This is said to be fixed, but I haven''t seen that myself, so I''m just wondering... I''ll get a 10TB test box released for testing OI in a few weeks, but before
2011 Jan 28
8
ZFS Dedup question
I created a zfs pool with dedup with the following settings: zpool create data c8t1d0 zfs create data/shared zfs set dedup=on data/shared The thing I was wondering about was it seems like ZFS only dedup at the file level and not the block. When I make multiple copies of a file to the store I see an increase in the deup ratio, but when I copy similar files the ratio stays at 1.00x. -- This