similar to: Re: Latest opensolaris VMware image

Displaying 20 results from an estimated 600 matches similar to: "Re: Latest opensolaris VMware image"

2008 Jan 30
2
vnic on top of aggr of 2 x 1 Gbit/s NICs
Hi, Can you tell why my vnic built on top of an aggregation of 2 NICs of 1 Gbit/s doesn''t report a speed of 2 GBit/s ? [root at nazgul /]# dladm show-dev bge0 link: up speed: 1000Mb duplex: full bge1 link: down speed: 0Mb duplex: unknown bge2 link: down speed: 0Mb duplex: unknown bge3 link: down speed: 0Mb
2006 Nov 16
6
DTrace hooks for CPU caps
Hello DTracers, The CPU caps project (http://www.opensolaris.org/os/project/rm/rctls/cpu-caps/) introduces kernel "wait queues" where threads may be placed to enforce caps. I would like to make this visible through DTrace and to add two new probes to the sched provider with the following semantics: cpucaps-sleep Probe that fires immediately before the current thread is placed on a
2003 Nov 12
5
repeat until function
Hi, I'm in this situation: I what to generate N random numbers(integer) that are different from each other. One suggestion: tabel <- rep(NULL, N) for (i in 1:N){ temp <- as.integer(runif(1,1,max)) if(temp in tabel) { repeat (?) (temp <- as.integer(runif(i,i,max))) until (?) ((temp in tabel) ==FALSE) } else{ tabel[i] <- temp} I can't use
2010 May 28
21
expand zfs for OpenSolaris running inside vm
hello, all I am have constraint disk space (only 8GB) while running os inside vm. Now i want to add more. It is easy to add for vm but how can i update fs in os? I cannot use autoexpand because it doesn''t implemented in my system: $ uname -a SunOS sopen 5.11 snv_111b i86pc i386 i86pc If it was 171 it would be grate, right? Doing following: o added new virtual HDD (it becomes
2008 May 29
1
>1TB ZFS thin provisioned partition prevents Opensolaris from booting.
Not sure where to put this but I am cc''ing the ZFS - discussion board. I was successfull in creating iscsi shares using ZFS set shareiscsi=on with 2 thin provisioned partitions of 1TB each (zfs create -s -V 1tb idrive/d1). Access to the shares with an iscsi initiator was successful, all was smooth, until the reboot. Upon reboot, the console reports the following errors. WARNING:
2010 Aug 10
4
[PATCH] Values of cpu_weight and cpu_cap are lost after xend restart
# HG changeset patch # User Lutz.Dube # Node ID ca2c8c262dbaa6b2521cec60af7cd688dabdf850 # Parent 8992134dcfd0b9e1e86f4111e68a8aa48bd33c3c Values of cpu_weight and cpu_cap are lost after xend restart For managed domains in state ''halted'' I always get default values for cpu_cap / cpu_weight after xend restart. This is because the names of parameters differ between a SXP file to
2009 Jan 14
1
RE: return codes ?
Hey I have just installed Samba 3.2.7 on my Debian 4.0 with compile. But I'm trying to to join my Windows AD, and samba returns Using short domain name -- UNDERVISNING Joined 'MAIL' to realm 'UNDERVISNING.LOCAL' return code = 0 First I got return code = -1 I just assumed that this was bad, so I powered on, but now I get return code = 0 but I still can't get any user
2006 Aug 17
4
[PATCH]configuration files(sched-credit)
Currently with the credit scheduler, cpu_weight cannot be assigned to guest domains at boot time.(it can be set with xm sched-weight after booting) This patch allows CPU weight, and CAP to be set by the Xen configuration files. Specification: The following can be added to the Xen configuration files. 1.CPU WEIGHT [OPTIONAL] cpu_weight = cpu weight value -sets the cpu time ratio to be
2008 Jun 21
12
Bfu xVM to build 92 problems
Bfu 91 to 92 looks good Proceed with xVM upgrade . After running:- # sunos.hg/bin/build-all nondebug bash-3.2# cd packages-nondebug bash-3.2# pwd /usr/tmp/packages-nondebug bash-3.2# ls -l total 20 drwxr-xr-x 4 root root 512 Jun 21 08:57 SUNWlibvirt drwxr-xr-x 4 root root 512 Jun 21 08:57 SUNWlibvirtr drwxr-xr-x 4 root root 512 Jun 21 08:57 SUNWurlgrabber
2009 Jan 13
12
OpenSolaris better Than Solaris10u6 with requards to ARECA Raid Card
Under Solaris 10 u6 , No matter how I configured my ARECA 1261ML Raid card I got errors on all drives that result from SCSI timeout errors. yoda:~ # tail -f /var/adm/messages Jan 9 11:03:47 yoda.asc.edu scsi: [ID 107833 kern.notice] Requested Block: 239683776 Error Block: 239683776 Jan 9 11:03:47 yoda.asc.edu scsi: [ID 107833 kern.notice] Vendor: Seagate
2006 Jul 31
3
LSI RAID Configuration Steps on T2000
Guru''s, Is anyone shar with me the steps to configure hardware RAID in T2000 server (LSI drivers) and use rootdisk hardware mirror -- Thanks a lot! Malahat Qureshi Ph.D. (MIS) Email: malahatq at hotmail.com -------------- next part -------------- An HTML attachment was scrubbed... URL:
2010 Nov 18
9
WarpDrive SLP-300
http://www.lsi.com/channel/about_channel/whatsnew/warpdrive_slp300/index.html Good stuff for ZFS. Fred -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://mail.opensolaris.org/pipermail/zfs-discuss/attachments/20101117/d48186f0/attachment.html>
2007 Sep 25
23
device alias
Hi. I''d like to request a feature be added to zfs. Currently, on SAN attached disk, zpool shows up with a big WWN for the disk. If ZFS (or the zpool command, in particular) had a text field for arbitrary information, it would be possible to add something that would indicate what LUN on what array the disk in question might be. This would make troubleshooting and general
2010 May 05
3
Another MPT issue - kernel crash
Hi all, I have faced yet another kernel panic that seems to be related to mpt driver. This time i was trying to add a new disk to a running system (snv_134) and this new disk was not being detected...following a tip i ran the lsitool to reset the bus and this lead to a system panic. MPT driver : BAD TRAP: type=e (#pf Page fault) rp=ffffff001fc98020 addr=4 occurred in module "mpt" due
2010 Aug 13
3
Guest CentOS statistic
Hi. I has xen virtual machine in paravirtualisation. Host and guest sysstems are CentOS. When I use top in guest machine I see, that system use 20% resourses. But in host machine I see, that machine use 50% resourses. In config machine has 5 processors, and has cpu_cap 352. Processor is 2266MHz. Therefor for statistic in guest machine it use ~ 2266 MHz(1 processor), and for Xen hypervisor - 4000
2007 Mar 10
4
Functions that cannot be traced by FBT under x86
Hi All: In Solaris Dtrace User Guide, page 218, it says, "Functions that do not create a stack frame on x86 systems cannot be instrumented by FBT." Then what are these functions? What characteristics do they have? Regards TJ
2007 Apr 19
14
Experience with Promise Tech. arrays/jbod''s?
Greetings, In looking for inexpensive JBOD and/or RAID solutions to use with ZFS, I''ve run across the recent "VTrak" SAS/SATA systems from Promise Technologies, specifically their E-class and J-class series: E310f FC-connected RAID: http://www.promise.com/product/product_detail_eng.asp?product_id=175 E310s SAS-connected RAID:
2011 Dec 21
8
Any rhyme or reason to disk dev names?
Hello, I am curious to know if there is an easy way to guess or identify the device names of disks. Previously the /dev/dsk/c0t0d0s0 system made sense to me... I had a SATA controller card with 8 ports, and they showed up with the numbers 1-8 in the "t" position of the device name. But I just built a new system with two LSI SAS HBAs in it, and my device names are along the lines of:
2009 Jan 15
21
4 disk raidz1 with 3 disks...
Hello, I was hoping that this would work: http://blogs.sun.com/zhangfan/entry/how_to_turn_a_mirror I have 4x(1TB) disks, one of which is filled with 800GB of data (that I cant delete/backup somewhere else) > root at FSK-Backup:~# zpool create -f ambry raidz1 c4t0d0 c5t0d0 c5t1d0 > /dev/lofi/1 > root at FSK-Backup:~# zpool list > NAME SIZE USED AVAIL CAP HEALTH ALTROOT
2006 Sep 29
4
[PATCH 4/6] xen: export NUMA topology in physinfo hcall
This patch modifies the physinfo hcall to export NUMA CPU and Memory topology information. The new physinfo hcall is integrated into libxc and xend (xm info specifically). Included in this patch is a minor tweak to xm-test''s xm info testcase. The new fields in xm info are: nr_nodes : 4 mem_chunks : node0:0x0000000000000000-0x0000000190000000