similar to: Major problem with a new ZFS setup

Displaying 20 results from an estimated 400 matches similar to: "Major problem with a new ZFS setup"

2013 Jun 25
1
Perplexed with environment
Hi I migrated from Linux to Mac, but I don't this has anything to do with it, but I am not sure. I am writing a small logger package, in which I have a file aaa.R: ,---- | .logData <- new.env() | assign("loggingThreshold", 10, envir = .logData) | assign("logToFile", FALSE, envir = .logData) | assign("logFileName", NULL, envir = .logData) | |
2001 Dec 14
1
Logistic regression : dicrepancies between glm and nls ?
Dear list, I'm trying to learn how to use nlme to be able to fit ad analyse mixed-model logistic regressions. In order to keep things simple, I started by the simplest possible model : a one (fixed-effect ...) continuous variable. This problem is, of course, solved by glm, but I wanted to look at a "hand-made" nls fit, in order to be able to "generalize" to nlme
2003 Jun 22
1
savecore: warning: /kernel version mismatch: ...
Morning all ... 'K, this one is a first for me ... server crashed this aft and savecore wouldn't dump the resultant core: pluto# savecore -v /vm/crash dumplo = 4362141696 (8519808 * 512) savecore: warning: /kernel version mismatch: "FreeBSD 4.8-STABLE #1: Sat May 31 22:57:04 ADT 2003 " and " #(#(#(" savecore: reboot savecore: dump time is zero
2010 Feb 16
3
converting character vector "hh:mm" to chron or strptime 24 clock time vectors
Hi All, I am attempting to work with some data from loggers. I have read in a .csv exported from MS Access that already has my dates and times (in 24 clock format), (with StringsAsFactors=FALSE). > head(tdata) LogData date time 1 77.16 2008/04/24 02:00 2 61.78 2008/04/24 04:00 3 75.44 2008/04/24 06:00 4 89.43 2008/04/24
2007 Jan 23
11
dynamic tablenames
Hi List! May I tap into your combined common sense ? I have a legacy logging app that needs to be modernized. At the moment, Data is going to be stored in 5 tables, one table is going to contain more than 200 million and 2 others about 60 million rows. Mysql will be used, unless someone sees a major advantage in using something else (at the moment data is stored in a raw positioned file format,
2003 Jun 25
0
savecore no longer works ... /kernel version mismatch
'k, now one of my other machines is getting the same thing on a crash: savecore: warning: /kernel version mismatch: "FreeBSD 4.8-STABLE #2: Fri Jun 20 18:34:14 ADT 2003 " and "" It was working, last successful savecore on that machine was Jun 6th: -rw------- 1 root wheel 4227792896 Jun 6 23:58 vmcore.1 -rw------- 1 root wheel 4227792896 Jun 6 10:37
2008 Jul 08
0
Disks errors not shown by zpool?
Ok, this is not a OpenSolaris question, but it is a Solaris and ZFS question. I have a pool with three mirrored vdevs. I just got an error message from FMD that read failed from one on the disks,(c1t6d0). All with instructions on how to handle the problem and replace the devices, so far everything is good. But the zpool still thinks everything is fine. Shouldn''t zpool also show
2012 Feb 04
2
zpool fails with panic in zio_ddt_free()
Hello all, I am not sure my original mail got through to the list (I haven''t received it back), so I attach it below. Anyhow, now I have a saved kernel crash dump of the system panicking when it tries to - I believe - deferred-release the corrupted deduped blocks which are no longer referenced by the userdata/blockpointer tree. As I previously wrote in my thread on unfixeable
2003 Jun 26
1
changes in kernel affecting savecore/dumps ...
David gave me some suggestions to check out on the servers, but so far, its all drawing a blank ... I have two servers right now that are updated to recent 4.8-STABLE kernels ... one was June 22nd, and the other was upgraded June 20th ... both of them have crashed since that date, and both of them tell me that they are unable to produce a core file, with the same errors: Jun 26 04:27:14 jupiter
2007 Jan 29
3
dumpadm and using dumpfile on zfs?
Hi All, I''d like to set up dumping to a file. This file is on a mirrored pool using zfs. It seems that the dump setup doesn''t work with zfs. This worked for both a standard UFS slice and a SVM mirror using zfs. Is there something that I''m doing wrong, or is this not yet supported on ZFS? Note this is Solaris 10 Update 3, but I don''t think that should
2004 Dec 02
2
Asterisk with SMS
Hi all, I am trying to setup the SMS of Asterisk. I have a Siemens SMS enable fixed phone which connects to my Asterisk through PSTN. Currently, I can use my fixed phone to edit and send messages to the Asterisk. However, I cannot make my Asterisk to send messages to the fixed phone. The SMS command displays TX and RX records, hang for a while and then stops with non-zero exits. I read
2011 Nov 25
1
Recovering from kernel panic / reboot cycle importing pool.
Yesterday morning I awoke to alerts from my SAN that one of my OS disks was faulty, FMA said it was in hardware failure. By the time I got to work (1.5 hours after the email) ALL of my pools were in a degraded state, and "tank" my primary pool had kicked in two hot spares because it was so discombobulated. ------------------- EMAIL ------------------- List of faulty resources:
2006 Apr 19
0
AHC Panic
I've finally been able to capture the panic, as now it occurs even with DDB configured. Of the six machines I have running 6.1-RC (CVSupped today), this is the only one that does this. /boot/kernel/kernel text=0x30c488 data=0x3b6a0+0x3170c syms=[0x4+0x46430+0x4+0x58da4] no such file or directory - Hit [Enter] to boot immediately, or any other key for command prompt. Booting
2003 Jul 02
0
union_lookup panics ...
grep union /var/log/messages Jul 2 12:53:01 jupiter savecore: reboot after panic: union_lookup returning . (0xc68e9e90) not same as startdir (0xc5e062c0) Jul 2 14:35:07 jupiter savecore: reboot after panic: union_lookup returning . (0xbf6fee90) not same as startdir (0xbb6d58c0) had two of them today, dumping nice cores ... I'm suspecting its someone trying to remove a file that is
2008 Aug 02
13
are these errors dangerous
Hi everyone, I''ve been running a zfs fileserver for about a month now (on snv_91) and it''s all working really well. I''m scrubbing once a week and nothing has come up as a problem yet. I''m a little worried as I''ve just noticed these messages in /var/adm/message and I don''t know if they''re bad or just informational: Aug 2 14:46:06
2005 Jul 18
1
scripts/create_db
I have been installing rails. I got through the //localhost:3000 information. At that point, I figured I need an app and have been working on salted hash login. I found http://wiki.rubyonrails.com/rails/show/SaltedLoginGeneratorQuickstart and worked through to where I want the create the db tables. Instead, I get script/create_db -bash: script/create_db: Permission denied [root@astarcc
2013 Feb 14
2
i386: vm.pmap kernel local race condition
Hi! I've got FreeBSD 8.3-STABLE/i386 server that can be reliably panicked using just 'squid -k rotatelog' command. It seems the system suffers from the problem described here: http://cxsecurity.com/issue/WLB-2010090156 I could not find any FreeBSD Security Advisory containing a fix. My server has 4G physical RAM (about 3.2G available) and runs squid (about 110M VSS) with 500
2007 Apr 23
3
ZFS panic caused by an exported zpool??
Apr 23 02:02:21 SERVER144 offline or reservation conflict Apr 23 02:02:21 SERVER144 scsi: [ID 107833 kern.warning] WARNING: /scsi_vhci/disk at g60001fe100118db00009119074440055 (sd82): Apr 23 02:02:21 SERVER144 i/o to invalid geometry Apr 23 02:02:21 SERVER144 scsi: [ID 107833 kern.warning] WARNING: /scsi_vhci/disk at g60001fe100118db00009119074440055 (sd82): Apr 23 02:02:21 SERVER144
2008 May 19
2
Suspend/resume on IBM X31
Greetings I am having trouble with suspend/resume on my Thinkpad X31, running 7.0-STABLE as of April 23. Any help would be appreciated. First problem: When I run "acpiconf -s3" from mulituser mode, the system suspends immediately, without executing /etc/rc.suspend (which has mode 755); then on resume, I get a panic. Second problem: When the system panics, I don't get a dump (or
2009 Jul 29
0
LVM and ZFS
I''m curious about if there are any potential problems with using LVM metadevices as ZFS zpool targets. I have a couple of situations where using a device directly by ZFS causes errors on the console about "Bus and lots of "stalled" I/O. But as soon as I wrap that device inside an LVM metadevice and then use it in the ZFS zpool things work perfectly fine and smoothly (no