similar to: problem with smbios 2.13?

Displaying 20 results from an estimated 100 matches similar to: "problem with smbios 2.13?"

2008 Oct 27
7
Fujitsu Siemens PRIMERGY RX300
Hi all Opensolaris works perfectly, but I am not able to boot the xvm kernel on this hardware. I added the -k option in grub, but the system hangs before the hostname line without any debug info. I''ve tried snv from b94 to b99, with the same results. If I install Debian with xen kernel I am able to use pvm and hvm guests. What can I do ? thanks Giacomo -- This message posted from
2007 Oct 31
6
b75 zone ip-exclusive problem
got a problem with b75 that the zone with ip-exclusive just takes the identity of the global zone. For the primary nic there is no option to set its ip and hostname. When i change the host name and ip the hostname does not change and the networking goes weird for both the zone and global zone. sparc eri qfe There any bugs about or should this work fine? This message posted from
2007 Nov 04
3
OpenSolaris for sparc in qemu on linux x64
Hi, i want to port an open-source project to sparc. I just downloaded OpenSolaris (sol-nv-b75-sparc-...iso), created an empty image for the hard disk, and tried to start qemu in a linux x64bit-host: qemu-system-sparc -hda test.img -cdrom sol-nv-b75-sparc-v1.iso -boot d -nographic The output says: Nvram id QEMU_BIOS, version 1 CPUs: 1 nvram error detected, zapping pram Welcome to OpenBIOS
2006 Jun 06
1
[Fwd: Persistent "Abort due to systemic unresponsiveness"]
Just adding a few more details to the previous mail so that I can avoid some trouble for you folks. 1. I have done a /execname != "dtrace"/ to avoid dtrace itself causing more syscalls to be fired. 2. I have also supressed the printing on the terminal which makes dtrace slower. And what happens is this: dtrace -qn ''syscall::read:entry /execname != "dtrace"/ {
2010 Jun 14
7
OpenSolaris Sparc Xen boot problem
I am running OpenSolaris build 134 on a T1000 (Sparc). It has been an uphill battle getting this far. Many of the commands I find in the various documentation won''t work for me. ------------------------------------------------------------ steve@t1000:~# pfexec pkg install xvm-gui Creating Plan pkg: ''xvm-gui'' matches multiple packages
2007 Oct 09
4
dom0 boot panic after bfu to b75
After BFU-ing my system to b75 I ended up with a panicing system when booting into Xen: grub> #---------- ADDED BY BOOTADM - DO NOT EDIT ---------- grub> title Solaris on Xen grub> kernel$ /boot/$ISADIR/xen.gz grub> module$ /platform/i86xpv/kernel/$ISADIR/unix /platform/i86xpv/kernel/$ISADIR/unix -B console=ttyb grub> module$ /platform/i86pc/$ISADIR/boot_archive grub>
2007 Oct 25
1
How to have ZFS root with /usr on a separate datapool
Ref: http://blogs.sun.com/timf/entry/zfs_bootable_datasets_happily_rumbling Ref: http://mediacast.sun.com/share/timf/zfs-actual-root-install.sh This is my errata for Tim Foster''s zfs root install script: 1/ Correct mode for /tmp should be 1777. 2/ The zfs boot install should allow you to have /usr on a separate zpool: a/ We need to create /zfsroot/usr/lib in the root partition and
2005 Nov 17
3
Abort due to systemic unresponsiveness
What is causing this error: dtrace: processing aborted: Abort due to systemic unresponsiveness It happens to my script which shows all the NFS operations coming through common_dispatch() on a server. I wasn''t the one running the test, so I don''t know how long it ran before it died. The script uses a couple of static probes and a couple fbt. Thanks, jim -------------- next
2007 Sep 19
3
ZFS panic when trying to import pool
I have a raid-z zfs filesystem with 3 disks. The disk was starting have read and write errors. The disks was so bad that I started to have trans_err. The server lock up and the server was reset. Then now when trying to import the pool the system panic. I installed the last Recommend on my Solaris U3 and also install the last Kernel patch (120011-14). But still when trying to do zpool import
2007 Nov 19
1
Recommended settings for dom0_mem when using zfs
I have a xVm b75 server and use zfs for storage (zfs root mirror and a raid-z2 datapool.) I see everywhere that it is recommended to have a lot of memory on a zfs file server... but I also need to relinquish a lot of my memory to be used by the domUs. What would a good value for dom0_mem on a box with 4 gig of ram?
2007 Nov 19
1
Recommended settings for dom0_mem when using zfs
I have a xVm b75 server and use zfs for storage (zfs root mirror and a raid-z2 datapool.) I see everywhere that it is recommended to have a lot of memory on a zfs file server... but I also need to relinquish a lot of my memory to be used by the domUs. What would a good value for dom0_mem on a box with 4 gig of ram?
2013 Apr 22
7
Multiple lon lat points in the map with ggplot2
Hello R users, For the last few days I am struggling with the following task: my data.frame: A1 A2 A3 B1 B2 B3 58.81 53.292 54.501 13.013 17.39 19.407 56.02 56.251 54.033 20.099 13.15 10.411 55.376 53.099 57.625 13.396 21.031 13.22 58.584 53.194 54.218 13.038 16.854 19.289 55.7 55.921 53.847 19.942 13.153 9.828 55.093 52.934
2008 Feb 08
16
Dom0 issues: snv_79b and Tecra M9
Hi all, I have a Toshiba Tecra M9 and have not been able to boot it dom0. This is running SXDE 01/08, snv79b. After booting under kmdb and setting moddebug=80000000 before booting the Solaris kernel (with help from Dan Mick), I was able to see mac_ether as the last thing loading, right after loading the e1000g driver. I cannot drop into kmdb via F1-A after it hangs. I''ve also
2013 May 28
1
Error "Illegal instruction (core dumped)" when running puppet agent (version 3.2.1)
Hi, I just upgraded my puppet to version to 3.2.1, and when running the agent, I get an error saying "Illegal instruction (core dumped)". Here is some details: - both master and agent is run on the same machine - # cat /etc/redhat-release CentOS release 6.2 (Final) - # uname -r 3.9.3-x86_64-linode33 - #ruby --version ruby 1.8.7 (2011-06-30 patchlevel 352) [x86_64-linux] -
2011 May 03
4
multipl disk failures cause zpool hang
Hi, There seems to be a few threads about zpool hang, do we have a workaround to resolve the hang issue without rebooting ? In my case, I have a pool with disks from external LUNs via a fiber cable. When the cable is unplugged while there is IO in the pool, All zpool related command hang (zpool status, zpool list, etc.), put the cable back does not solve the problem. Eventually, I
2008 Jun 21
12
Bfu xVM to build 92 problems
Bfu 91 to 92 looks good Proceed with xVM upgrade . After running:- # sunos.hg/bin/build-all nondebug bash-3.2# cd packages-nondebug bash-3.2# pwd /usr/tmp/packages-nondebug bash-3.2# ls -l total 20 drwxr-xr-x 4 root root 512 Jun 21 08:57 SUNWlibvirt drwxr-xr-x 4 root root 512 Jun 21 08:57 SUNWlibvirtr drwxr-xr-x 4 root root 512 Jun 21 08:57 SUNWurlgrabber
2007 Dec 14
0
scrub percentage complete decreasing, but without snaps.
I''ve seen the problems with bug 6343667, but I haven''t seen the problem I have at the the moment. I started a scrub of a b72 system that doesn''t have any recent snapshots (none since the last scrub) and the % complete is cycling: scrub: scrub in progress, 69.08% done, 0h13m to go scrub: scrub in progress, 46.63% done, 0h28m to go scrub: scrub in progress, 6.36%
2011 Feb 16
2
Regenerating mdbox/sis
It's apparent I have some level of corruption in my mailboxes - something to do with the combination of mdbox, sis, and zlib. I should note that I know I have a mix of compressed and raw messages - if that makes any difference in identifying the problems. I think I need to somehow regenerate the mailboxes with issues. Is there a suggestion for how to accomplish this? Should I perhaps use
2007 Sep 28
5
ZFS Boot Won''t work with a straight or mirror zfsroot
Using build 70, I followed the zfsboot instructions at http:// www.opensolaris.org/os/community/zfs/boot/zfsboot-manual/ to the letter. I tried first with a mirror zfsroot, when I try to boot to zfsboot the screen is flooded with "init(1M) exited on fatal signal 9" Than I tried with a simple zfs pool (not mirrored) and it just reboots right away. If I try to setup grub
2015 Dec 21
3
lldb -c corefile get segmentation fault on centos7
Hi, I build llvm+clang+lldb 3.7 successfully on centos7, and lldb -p PID works pretty well. However when I tried lldb -c corefile executable_bin, lldb itself core dumpped. Attached the following core info which is debugged by gdb: [root at dn-cn-controller-4fbd4 data1]# lldb -c a.corefile /usr/local/myproject/bin/cnode *(lldb) target create "/usr/local/myproject/bin/cnode" --core