search for: 1.7gb

Displaying 20 results from an estimated 25 matches for "1.7gb".

Did you mean: 1.7g
2011 Jul 20
3
Memory Usage in Top and System Monitor
Can someone help clear up some confusion in reading memory usage in Top and System monitor. Here is a picture of both. http://s1176.photobucket.com/albums/x327/ionosphere2011/ Why does "System Monitor" show 1.7gb free out of 8gb. While Top shows all 8gb being used? If 1.7gb is free then it should not be using swap space so I assume "System Monitor" is reporting it
2010 Jul 28
2
memory problem for scatterplot using ggplot
Dear all, I have a memory problem in making a scatter plot of my 17.5 million-pair datasets. My intention to use the "ggplot" package and use the "bin2d". Please find the attached script for more details. Could somebody please give me any clues or tips to solve my problem?? please ... Just for additional information: I'm running my R script on my 32-bit machine: Ubuntu
2016 Nov 16
2
Re: [ovirt-users] OVA import of FC21 VM hangs during virt-v2v conversion?
Hi, On Wed, November 16, 2016 5:15 pm, Richard W.M. Jones wrote: > On Wed, Nov 16, 2016 at 05:09:56PM -0500, Derek Atkins wrote: > > I'll try to reproduce the issue here, but you can also do > the following command directly on the guest disk image if you > want to test something: > > time LIBGUESTFS_BACKEND=direct guestfish --ro -a fc21-64.qcow2 -i > selinux-relabel
2007 Apr 15
4
Ferret and Mongrel. OSX vs. Linux
I''m having a strange problem accessing a 1.7GB Ferret index from within Mongrel (1.0.1) on Linux. On OSX a Ferret search through Rails takes a fraction of a second. From the command line, bypassing Mongrel, the search takes about the same amount of time. On Fedora Core 4 a Ferret search from the command line takes a fraction of a second, but the same search through Mongrel never
2007 Apr 15
4
Ferret and Mongrel. OSX vs. Linux
I''m having a strange problem accessing a 1.7GB Ferret index from within Mongrel (1.0.1) on Linux. On OSX a Ferret search through Rails takes a fraction of a second. From the command line, bypassing Mongrel, the search takes about the same amount of time. On Fedora Core 4 a Ferret search from the command line takes a fraction of a second, but the same search through Mongrel never
2008 Feb 18
2
predicting memory usage
Hi All, is there a way of predicting memory usage? I need to build an array of 86000 by 2500 numbers (or I might create a list of 2 by 2500 arrays 43000 long). How much memory should I expect to use/need? Cheers, Fede -- Federico C. F. Calboli Department of Epidemiology and Public Health Imperial College, St. Mary's Campus Norfolk Place, London W2 1PG Tel +44 (0)20 75941602 Fax +44
2014 Dec 03
4
[LLVMdev] FAQ update and Question on minimum build
Sorry to trouble you, great software, I have it on Ubuntu 14.10, Windows 10 and now I was trying ARM: I have been trying to build a working Cland+LLVM for 4 days now, ARM Radxa Rock Pro Quad 1.6Ghz 2GB memory I'm building on an SD card, I tried 4GB and died(space thing), 8GB and died(again with the space thing) so I read everything I could find on building only a minimum testing build and I
2016 Nov 16
0
Re: [ovirt-users] OVA import of FC21 VM hangs during virt-v2v conversion?
On Wed, Nov 16, 2016 at 05:30:55PM -0500, Derek Atkins wrote: > Hi, > > On Wed, November 16, 2016 5:15 pm, Richard W.M. Jones wrote: > > On Wed, Nov 16, 2016 at 05:09:56PM -0500, Derek Atkins wrote: > > > > I'll try to reproduce the issue here, but you can also do > > the following command directly on the guest disk image if you > > want to test
2007 Apr 15
3
Massive difference in query times on OSX and Linux
I have a 1.7GB index with 440818 documents. Queries that take 0.23846 seconds on my MacBook pro take an indefinite period of time on a Fedora Core 4 EC2 instance. Five minutes and counting...I''m using Ferret 0.11.3. I don''t see any errors in the logs. Has anyone ever seen anything like this?
2006 Jun 15
1
Rsync memory usage seems a lot higher than 100bytes/file
All, I have read the lists, I have read the faq - yes, rsync uses a lot of memory. That said - I need to use rsync in an environment with a lot of files, and I need to use -H and --delete options. The faq says about 100 bytes per file - at the moment I am looking at 8700000 files, which should translate to about 800MB - but it does not. In fact, running rsync not only consumes all of the ram,
2016 Nov 17
2
Re: [ovirt-users] OVA import of FC21 VM hangs during virt-v2v conversion?
Hi, On Wed, November 16, 2016 5:39 pm, Richard W.M. Jones wrote: > On Wed, Nov 16, 2016 at 05:30:55PM -0500, Derek Atkins wrote: >> Hi, >> >> On Wed, November 16, 2016 5:15 pm, Richard W.M. Jones wrote: >> > On Wed, Nov 16, 2016 at 05:09:56PM -0500, Derek Atkins wrote: >> > >> > I'll try to reproduce the issue here, but you can also do >>
2006 Jan 24
2
importing a VERY LARGE database from Microsoft SQL into R
I am using R 2.1.1 in a Windows Xp environment. I need to import a large database from Microsoft SQL into R. I am currently using the ?sqlQuery? function/command. This works, but I sometimes run out of memory if my database is too big, or it take quite a long time for the data to import into R. Is there a better way to bring a large SQL database into R? IS there an efficient way to convert
2007 May 18
5
how to compile with large file support?
Hi, I''m trying to figure out how to compile ferret with large file support, but none of the topics that discuss this actually say How this is done. Can someone please provide the info? thanks. -m my exact problem: http://www.ruby-forum.com/topic/94143#191630 this topic also discusses the issue: http://www.ruby-forum.com/topic/84237#151791 this topic says that the FAQ should have the
2016 Nov 16
2
Re: [ovirt-users] OVA import of FC21 VM hangs during virt-v2v conversion?
Hi, On Wed, November 16, 2016 5:03 pm, Richard W.M. Jones wrote: > On Wed, Nov 16, 2016 at 04:49:46PM -0500, Derek Atkins wrote: >> I suppose there could be a bug. I just verified that it's absolutely >> setfiles: >> >> 14:07:25 E: commandrvf: setfiles -F -e /sysroot/dev -e /sysroot/proc -e >> /sysroot/selinux -e /sysroot/sys -r /sysroot -q >>
2007 May 25
2
R scripts slowing down after repeated called to compiled code
Thanks in advance to anyone that might be able to help me with this problem. I have not been able to find a reference to it in the documentation on online sources, so I am turning to this group. I am running R 2.4.1 under Red Hat Enterprise Linux 4, on an x86_64 platform (multi-core Intel Xeon processors, 3.6Ghx, 8GB of RAM). I have some rather complicated code (so I won't attach it here),
2014 Oct 14
2
[LLVMdev] [RFC] Less memory and greater maintainability for debug info IR
> On Oct 13, 2014, at 6:59 PM, Sean Silva <chisophugis at gmail.com> wrote: > > Stupid question, but when I was working on LTO last Summer the primary culprit for excessive memory use was due to us not being smart when linking the IR together (Espindola would know more details). Do we still have that problem? For starters, how does the memory usage of just llvm-link compare to the
2003 Oct 09
3
Specifying suitable PC to run R
If I am buying a PC where the most compute intensive task will be running R and I do not have unlimited resources what trade-offs should I make? Specifically should I go for 1 - more memory, or 2 - faster processor, or 3 - something else? If it makes a difference I shall be running Windows on it and I am thinking about getting a portable which I understand makes upgrading more difficult. Extra
2016 Jul 20
2
[XRay] Build instrumented Clang, some analysis results
Hi everyone, TL;DR: With current pending patches applied in compiler-rt and llvm, and trunk clang, you can build your application with XRay tracing enabled on Linux with tracing enabled before main starts, and logging stops when the main thread exits. Just a quick update, I have some patches under review that when applied cleanly to LLVM and compiler-rt allows for building applications with XRay
2003 Nov 19
11
Windows R 1.8.0 hangs when Mem Usage >1.8GB
I have a loop that increases the size of an object after each iteration. When the Windows Task Manager shows "Mem Usage" about 1.8GB, the Rgui.exe process no longer responds. I use: "C:\Program Files\R\rw1080\bin\Rgui.exe" --max-mem-size=4000M --min-vsize=10M --max-vsize=3000M --min-nsize=500k --max-nsize=1000M I have a dual Xeon 2.8GHz processor box with 4GB of memory and
2002 Oct 18
7
RAM usage
Hi, I'm having problems while working with large data sets with R 1.5.1 in windows 2000. Given a integer matrix size of 30 columns and 15000 rows my function should return a boolean matrix size of about 5000 rows and 15000 columns. First of all I tried to run this function on computer with 256 MB of RAM. I increased memory limit of R with memory.limit() up to 512 MB. I was inspecting