similar to: Newbie questions -- is OCFS2 what I even want?

Displaying 20 results from an estimated 400 matches similar to: "Newbie questions -- is OCFS2 what I even want?"

2009 Nov 13
1
Cannot set heartbeat dead threshold
Hi I have: SLES 10 SP2 (2.6.16.60-0.21-smp) ocfs2-tools-1.4.0-0.3 ocfs2console-1.4.0-0.3 and I can't change "heartbeat dead threshold" value. Content of /etc/sysconfig/o2cb: # O2CB_ENABLED: 'true' means to load the driver on boot. O2CB_ENABLED=true # O2CB_BOOTCLUSTER: If not empty, the name of a cluster to start. O2CB_BOOTCLUSTER=ocfs2 # O2CB_HEARTBEAT_THRESHOLD:
2006 May 26
1
Another node is heartbeating in our slot!
All, We are having some problems getting OCFS2 to run, we are using kernel 2.6.15 with OCFS2 1.2.1. Compiling the OCFS2 sources went fine and all modules load perfectly. However, we can only mount the OCFS2 volume on one machine at a time, when we try to mount the volume on the 2 other machines we get an error stating that another node is heartbeating in our slot. When we mount the volume
2010 Jan 18
1
Getting Closer (was: Fencing options)
One more follow on, The combination of kernel.panic=60 and kernel.printk=7 4 1 7 seems to have netted the culrptit: E01-netconsole.log:Jan 18 09:45:10 E01 (10,0):o2hb_write_timeout:137 ERROR: Heartbeat write timeout to device dm-12 after 60000 milliseconds E01-netconsole.log:Jan 18 09:45:10 E01 (10,0):o2hb_stop_all_regions:1517 ERROR: stopping heartbeat on all active regions.
2011 Mar 03
1
OCFS2 1.4 + DRBD + iSCSI problem with DLM
An HTML attachment was scrubbed... URL: http://oss.oracle.com/pipermail/ocfs2-users/attachments/20110303/0fbefee6/attachment.html
2008 Aug 21
5
VM node won't talk to host
I am trying to mount the same partition from a KVM ubuntu 8.04.1 virtual machine and on an ubuntu 8.04.1 host server. I am able to mount the partition just on fine on two ubuntu host servers, they both talk to each other. The logs on both servers show the other machine mounting and unmounting the drive. However, when I mount the drive in the KVM VM I get no communication to the host
2008 Oct 22
2
Another node is heartbeating in our slot! errors with LUN removal/addition
Greetings, Last night I manually unpresented and deleted a LUN (a SAN snapshot) that was presented to one node in a four node RAC environment running OCFS2 v1.4.1-1. The system then rebooted with the following error: Oct 21 16:45:34 ausracdb03 kernel: (27,1):o2hb_write_timeout:166 ERROR: Heartbeat write timeout to device dm-24 after 120000 milliseconds Oct 21 16:45:34 ausracdb03 kernel:
2008 Sep 18
2
o2hb_do_disk_heartbeat:982:ERROR
Hi everyone; I have a problem on my 10 nodes cluster with ocfs2 1.2.9 and the OS is RHEL 4.7 AS. 9 nodes can start o2cb service and mount san disks on startup however one node can not do that. My cluster configuration is : node: ip_port = 7777 ip_address = 192.168.5.1 number = 0 name = fa01 cluster = ocfs2 node: ip_port =
2004 Dec 26
3
Help building online game
Now I just need some helpful hints from the community as to what I could use to display a hexagonal playing field of about 40x40 hex units. Trying to build an online strategy game with Ruby on Rails. Working out the methodology on paper and then scaffolding some. Any ideas which technologies I could possibly use ? other than just simple .gifs as icons and maybe overlayed css graphics ? Would
2010 May 31
2
O2CB_HEARTBEAT_THRESHOLD won't take changes
Hello All, I have multiple OCFS2 clusters on SLES10 SP2 running Xen. We needed to increase the O2CB_HEARTBEAT_THRESHOLD from 31 up to 61 and did so successfully on 2 of our 3 clusters. However on one of the three clusters we are not able to change the value. The /etc/sysconfig/o2cb file contains 61 as the threshold after reconfiguring via /etc/init.d/o2cb configure, we reconfigure all 3 nodes at
2003 Jan 10
5
The community
It occurred to me over the last few days that in fact I hadn''t been a model ShoreWall Citizen so I went to the site and visited this link for inspiration: http://www.cityofshoreline.com/about/history/index.html Here I was reminded of just how open source communities are suppose to work. Just like the early beginnings of Shoreline, Tom has brought together many scattered settlements
2007 Jul 25
4
Problem installing on RH3 U8
Hi, i dont seem to be able to get ocfs running on RH3 U8 32Bit [root@libra-devb-db1 root]# uname -a Linux devb-db1.mydomain 2.4.21-47.ELsmp #1 SMP Wed Jul 5 20:38:41 EDT 2006 i686 athlon i386 GNU/Linux [root@devb-db1 root]# cat /etc/redhat-release Red Hat Enterprise Linux AS release 3 (Taroon Update 8) [root@devb-db1 root]# rpm -ivh ocfs-2.4.21-EL-smp-1.0.14-1.i686.rpm Preparing...
2006 Nov 21
2
Memory leak in ocfs2/dlm?
Hi! Seems we're facing some memory leak here. This is vanilla 2.6.19-rc6 on a x86_64 box, 4GB RAM. A simple `ls -Rn' on a filesystem with lots of files makes the box leak so much RAM that the OOM killer starts to kick in. With slab alloc debugging turned on, we see this: # mount; ls -Rn; wait some seconds; Ctrl-C [root@lnxp-1038:/backend1]$ cat /proc/slab_allocators | egrep
1997 Sep 29
1
UMASK Problems
I have quite a problem with the new SAMBA version(s) ( both 1.9.17 and 1.9.17 p2). When creating a directory in a samba-mounted unix file system the create mask entry will be ignored. I tried a lot (changing umasks on the unix boxes (SunOS/Solaris), changing the create mask entry to a many different values) but nothing happens at all. Every directory will be created with the permissions of 755.
2010 Oct 23
1
Reg: ocfs2 two node cluster crashed, node2 crashed, when I rebooted node1 for maintenance.
Hi All, We have ocfs2 node cluster with oracle 11G RAC running, The node2 got crashed automatically, when i rebooted node one for maintenance. please check the log from node2 , before its got crashed. Oct 23 15:42:25 node2 kernel: ocfs2_dlm: Nodes in domain ("029C02C993E44E90879922E268FB161A"): 2 Oct 23 15:42:29 node2 kernel: ocfs2_dlm: Node 1 leaves domain
2009 Jul 29
3
Error message whil booting system
Hi, When system booting getting error message "modprobe: FATAL: Module ocfs2_stackglue not found" in message. Some nodes reboot without any error message. ------------------------------------------------- ul 27 10:02:19 alf3 kernel: ip_tables: (C) 2000-2006 Netfilter Core Team Jul 27 10:02:19 alf3 kernel: Netfilter messages via NETLINK v0.30. Jul 27 10:02:19 alf3 kernel:
2007 Aug 22
1
mount.ocfs2: Value too large ...
Hallo, I have two servers and both are connected to external array, each by own SAS connection. I need these servers to work simultaneously with data on array and I think that ocfs2 is suitable for this purpose. One server is P4 Xeon (Gentoo linux, i386, 2.6.22-r2) and second is Opteron (Gentoo linux, x86_64, 2.6.22-r2). Servers are connected by ethernet, adapters are both Intel
2007 Mar 16
2
re: o2hb_do_disk_heartbeat:963 ERROR: Device "sdb1" another node is heartbeating in our slot!
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Folks, I'm trying to wrap my head around something that happened in our environment. Basically, we noticed the error in /var/log/messages with no other errors. "Mar 16 13:38:02 dbo3 kernel: (3712,3):o2hb_do_disk_heartbeat:963 ERROR: Device "sdb1": another node is heartbeating in our slot!" Usually there are a
2009 Oct 20
1
Fencing OCFS
I people, i install the ocfs in my virtual machine, with centos 5.3 and xen. But when i turn off the machine1, the ocfs start the fencing off the machine2. I read the doc in oracle.com, but could not solve the problem. Someone help? My conf and my package version. cluster.conf node: ip_port = 7777 ip_address = 192.168.1.35 number = 0 name = x1 cluster
2006 Aug 01
1
AW: ocfs2_search_chain: Group Descriptor has bad signature
I'm using ocfs2 and all modules from Suse (SLES9), no self compilations. Here are the details: * 32-bit machine (writing to ocfs2 partition/LUN and where the corruption was reported): Kernel: 2.6.5-7.257-bigsmp #1 SMP i686 i386 GNU/Linux OCFS2 rpms: ocfs2console-1.2.1-4.2 ocfs2-tools-1.2.1-4.2 o2cb_ctl -V: o2cb_ctl version 1.2.1 /etc/init.d/o2cb status: Module "configfs":
2011 Jul 06
2
Slow umounts on SLES10 patchlevel 3 ocfs2
Hi, we are using a SLES10 Patchlevel 3 with 12 Nodes hosting tomcat application servers. The cluster was running some time (about 200 days) without problems. Recently we needed to shutdown the cluster for maintenance and experianced very long times for the umount of the filesystem. It took something like 45 minutes each node and filesystem (12 x 45 minutes shutdown time). As a result the planned