similar to: Init array to -1 with memset()?

Displaying 20 results from an estimated 1000 matches similar to: "Init array to -1 with memset()?"

2003 Oct 05
2
Possible security hole
Maybe security related mails should be sent elsewhere? I didn't notice any so here it goes: sender.c:receive_sums() s->count = read_int(f); .. s->sums = (struct sum_buf *)malloc(sizeof(s->sums[0])*s->count); if (!s->sums) out_of_memory("receive_sums"); for (i=0; i < (int) s->count;i++) { s->sums[i].sum1 = read_int(f);
2004 Jan 26
1
How match.c hash_search works with multiple checksums that have identical tags
I am trying to understand how match.c works. I am reading the code and something doesnt look quite right. This is usually a sign that I am missing something obvious. Here is what I see. build_hash_table uses qsort to order targets in ascending order of //tag,index// into the array of checksums. It then accesses the targets in ascending order and writes the index at the tag's location in
2018 Jun 15
2
[WebAssembly] lld dynamic loader
Hi Sam, Thanks for your answer! Sorry I didn't really know how to process, but this issue block me to make some legacy C code works without having some glue with a bit static lib ;( (The change have really low risk) The changes I used locally to make dlopen/dlclose et dlsym mechanism works :) : Index: wasm/Writer.cpp =================================================================== ---
2018 Jun 01
2
[WebAssembly] lld dynamic loader
Hello, The table generated by lld (Wasm target) have a max size. So this make impossible to add some others functions in the table . As for the moment only one Table is available in WAsm this is the only way to add function in a dynamic way. Having a none static size of the table is useful for a kind dynamic loader/inker at run-time: by adding some exported functions coming from another
2012 Feb 16
2
[LLVMdev] LLVM GHC Backend: Tables Next To Code
> Since the jmp is a fixed 2 bytes (0xEB, tablesize), all references to the table can still be done with trivial pc/RIP-relative addressing within the closure, and you just need one pointer for both the table and the closure data. Only if tablesize fits in 127 bytes. Otherwise it will use 0xE9-based encoding which is larger. -- With best regards, Anton Korobeynikov Faculty of Mathematics and
2012 Feb 16
1
[LLVMdev] LLVM GHC Backend: Tables Next To Code
> > Since the jmp is a fixed 2 bytes (0xEB, tablesize), all references to > the table can still be done with trivial pc/RIP-relative addressing within > the closure, and you just need one pointer for both the table and the > closure data. > Only if tablesize fits in 127 bytes. Otherwise it will use 0xE9-based > encoding which is larger. The table size is variable - between 2
2002 Apr 19
2
out of memory in build_hash_table
I recently installed rsync 2.5.5 on both my rsync server and client being used. I installed the latest version because I was having problems with rsync stalling with version 2.4.6 (I read that 2.5.5 was supposed to clear this up or at least give more appropriate errors). I am still having problems with rsync stalling even after upgrading to 2.5.5. It only stalls in the "/home" tree
2002 Mar 23
1
why variable last_i is needed in match.c rsync source ?
Hi all I see the rsync source and rsync makes hashing table and search hashing table tag_table to find the index of array struct sum_buf , which is a element of struct sum_struct. According to the source code, variable last_i is used to encourage adjacent matches allowing the RLL coding of the output to work more efficiently. Why last_i makes more efficiency? I can't understanding what
2012 Feb 15
0
[LLVMdev] LLVM GHC Backend: Tables Next To Code
On Feb 14, 2012, at 10:30 AM, David Terei wrote: > Hmm writing a blog post about TNTC is beyond the time I have right now. Sure, understandable. I'm surprised someone else hasn't already :) > Here is some high level documentation of the layout of Heap objects in GHC: > > http://hackage.haskell.org/trac/ghc/wiki/Commentary/Rts/Storage/HeapObjects#InfoTables > > With
2006 Mar 23
1
AIX 5.1 rsync large file
Hello, I have inherited a setup where there are 2 AIX 5.1 systems in 2 separate sites. There are large database files that are backed from each site to the other via rsync. Currently, it is using rsync version 2.5.4. It does it via ssh with the options -avz. This has all been merrily plodding along for some time. There is one file that is over 45 GB, and it started having trouble with that
2004 Jun 19
1
[Bug 1467] Hash table generation seems to be flawed
https://bugzilla.samba.org/show_bug.cgi?id=1467 ------- Additional Comments From ripper@internode.on.net 2004-06-19 02:18 ------- The way the code seems to work is as follows: A sorted list of hashes and their block numbers are generated (targets). They're sorted so that identical hashes are numerically next to each other. A reverse lookup table is then generated, that (i presume) is
2008 Dec 09
1
expected variable name error pos 98349 WInBUGS in R
> I am using a random intercept model with SITEID as random and NAUSEA as > outcome. > > I tried using a dataset without missing values and changed my model > statement accordingly but still get the same error. Follwoing in an excerpt. > > anal.data <- read.table("nausea.txt", header=T, sep="\t") > > list(names(anal.data)) > [[1]] > [1]
2002 Aug 05
5
[patch] read-devices
Greetings, I'd like to propose a new option to rsync, which causes it to read device files as if they were regular files. This includes pipes, character devices and block devices (I'm not sure about sockets). The main motivation is cases where you need to synchronize a large amount of data that is not available as regular files, as in the following scenarios: * Keep a copy of a block
2008 Dec 07
1
Reading txt file in R
Hi:   I am using the following code to read a data set in txt in R and using the follwoing model. But it seems to give me an error ' expected variable name error pos 134022'. Any help is greatly appreciated.   Code: anal.data <- read.table("nausea.txt", header=T, sep="\t") list(names(anal.data)) attach(anal.data) n.samples <- dim(anal.data) [1] # number of data
2008 Dec 08
1
Reading txt file in R to run Random Intercept Model
I am using a random intercept model with SITEID as random and NAUSEA as outcome. Thanks. Anamika I tried using a dataset without missing values and changed my model statement accordingly but still get the same error. Follwoing in an excerpt. > anal.data <- read.table("nausea.txt", header=T, sep="\t") > list(names(anal.data)) [[1]] [1] "SITEID"
2011 Oct 27
2
Syntax Check: rshape2 melt()
This is my first excursion into using reshape2 and I want to ensure that the melt() function call is syntactically correct. The unmodifed data frame is organized this way: head(tds.anal) site sampdate param quant 1 UDS-O 2006-12-06 TDS 10800 4 STC-FS 1996-06-14 Cond 280 7 UDS-O 2007-10-04 Mg 1620 9 UDS-O 2007-10-04 SO4 7580 19 JCM-10B 2007-06-21 Ca 79 20
2011 Oct 24
2
Syntax Help for xyplot()
Thanks to David's help I subset my large data set and produced a smaller one for a single stream and 7 factors of interest. The structure of this data frame is: str(burns.tds.anal) 'data.frame': 718 obs. of 4 variables: $ site : Factor w/ 143 levels "BC-0.5","BC-1",..: 1 1 4 6 4 4 4 5 5 5 $ sampdate: Date, format: "1996-06-02"
2012 Feb 15
2
[LLVMdev] LLVM GHC Backend: Tables Next To Code
On Feb 15, 2012, at 12:16 PM, Chris Lattner <clattner at apple.com> wrote: > > On Feb 14, 2012, at 10:30 AM, David Terei wrote: > >> Hmm writing a blog post about TNTC is beyond the time I have right now. > > Sure, understandable. I'm surprised someone else hasn't already :) > >> Here is some high level documentation of the layout of Heap objects
2012 Feb 15
0
[LLVMdev] LLVM GHC Backend: Tables Next To Code
> This is starting to look very similar to how ARM constant islands work, without the extra ugliness from how small the ARM immediate displacements are. > > -Jim Would there be any reason that this couldn't be seen as an opportunity to move the constant islands pass out of the ARM backend and make the target-independent constant pools (which ARM bypasses completely) more generic?
2013 Feb 21
1
Ideal RAID-1 Stripe Size for IMAP Server ?
Hi there. I'm currently configuring a brand new email server(IMAP) which will run CentOS 6.3 x64. So lots of files on the data volume. I'm confused about stripe size for my data volume. I'm using 2x500GB for OS in RAID1 and 2x2TB SAS for DATA in RAID1 which is hosted on a HP Smart Array P420 Controller with 1GB in Cache/Battery Backed Up. Both these array's are on the P420