similar to: Adding new nodes to OCFS2?

Displaying 20 results from an estimated 200 matches similar to: "Adding new nodes to OCFS2?"

2006 Apr 08
4
source routing does not work with extra ip addresses
I set up this config: +------+ -+ ISP1 +--+ +------+ | +-------+ +--+ linux | +------+ | +-------+ -+ ISP2 +--+ +------+ No problem. Standard setup with two ISP''s. Both routed subnets. Default gateway is ISP1. No magic here. Now I put a server behind the Linux box. I want the server to be reachable on an /extra/ IP in the routed subnet of ISP2. +------+ -+ ISP1
2005 Dec 30
1
NewbieQ concatenating name in options_from_collection_for_select
Hi, A real dumb one :~/ How do you concatenate a first_name, last_name as the text method for OFCFS? Damned if I can do it without error. Is it in the original select for the collection? Eric.
2008 Sep 18
2
o2hb_do_disk_heartbeat:982:ERROR
Hi everyone; I have a problem on my 10 nodes cluster with ocfs2 1.2.9 and the OS is RHEL 4.7 AS. 9 nodes can start o2cb service and mount san disks on startup however one node can not do that. My cluster configuration is : node: ip_port = 7777 ip_address = 192.168.5.1 number = 0 name = fa01 cluster = ocfs2 node: ip_port =
2006 Jan 08
1
Options_from_collection_for_select -> Multiple fields in display
The following code shows a list of the users, with username being each person''s initials (mine would be DM). <%= options_from_collection_for_select @users,"id","username" %> However, I''d like to have the display of the SELECT box be "DM - Dylan Markow" instead of "DM." Is this possible using
2007 Sep 06
1
60% full and writes fail..
I have a setup with lot's of small files (Maildir), in 4 different volumes and for some reason the volumes are full when they reach 60% usage (as reported by df ). This was ofcourse a bit of a supprise for me .. lots of failed writes, bounced messages and very angry customers. Has anybody on this list seen this before (not the angry customers ;-) ? Regards, =paulv # echo "ls
2008 Oct 03
2
OCFS2 with Loop device
hi there i try to setup OCFS2 with loop device /dev/loop0 i've 4 servers running SLES10 SP2. internal ip's: 192.168.55.1, .2, .3 and .6 my cluster.conf: -------------------------------------------- node: ip_port = 7777 ip_address = 192.168.55.1 number = 0 name = www cluster = cawww node: ip_port = 7777 ip_address = 192.168.55.2
2011 Jun 27
1
multiple cluster doesn't work
We're trying to setup 3 PRDM partitions (VMware) across 2 nodes. As long as only one is configured in cluster.conf, there's not problem. As soon as we try to use 2 or more we get issues. It looks the same as bug 636: http://oss.oracle.com/bugzilla/show_bug.cgi?id=636 I posted my cluster.conf and command line results there. I'm including them here in the hopes that someone on this
2008 Sep 10
4
mount.ocfs2: Error when attempting to run /sbin/ocfs2_hb_ctl: "Operation not permitted".
Hi, I am trying to configure a two node cluster on SLES10SP2 using user level heartbeat. Here is my configuration. ocfs2-tools-1.4.0-0.3 **user level heartbeat** -> lsmod | grep ocfs ocfs2_user_heartbeat 20992 1 ocfs2_dlmfs 37776 1 ocfs2_dlm 204456 1 ocfs2_dlmfs ocfs2_nodemanager 223384 6 ocfs2_user_heartbeat,ocfs2_dlmfs,ocfs2_dlm configfs 44700 3 ocfs2_user_heartbeat,ocfs2_nodemanager
2008 Apr 23
2
Transport endpoint is not connected
Hi, I want to share a partition of an external disk between 3 machines. It is located on a Sun StoreTek 3510 and bound (via multipathing) to all 3 machines. One machine (say mach1) is the node I used to create the filesystem and setup /etc/ocfs2/cluster.conf. I copied the file to the other 2 machines. All use the same /etc/multipath.conf[1], too. All start /etc/init.d/ocfs2 at boot, all are
2007 Sep 30
4
Question about increasing node slots
We have a test 10gR2 RAC cluster using ocfs2 filesystems for the Clusterware files and the Database files. We need to increase the node slots to accomodate new RAC nodes. Is it true that we will need to umount these filesystems for the upgrade (i.e. Database and Clusterware also)? We are planning to use the following command format to perform the node slot increase: # tunefs.ocfs2 ?N 3
2011 Mar 03
1
OCFS2 1.4 + DRBD + iSCSI problem with DLM
An HTML attachment was scrubbed... URL: http://oss.oracle.com/pipermail/ocfs2-users/attachments/20110303/0fbefee6/attachment.html
2008 Mar 05
3
cluster with 2 nodes - heartbeat problem fencing
Hi to all, this is My first time on this mailinglist. I have a problem with Ocfs2 on Debian etch 4.0 I'd like when a node go down or freeze without unmount the ocfs2 partition the heartbeat not fence the server that work well ( kernel panic ). I'd like disable or heartbeat or fencing. So we can work also with only 1 node. Thanks
2011 Dec 06
2
OCFS2 showing "No space left on device" on a device with free space
Hi , I am getting the error "No space left on device" on a device with free space which is ocfs2 filesystem. Additional information is as below, [root at sai93 staging]# debugfs.ocfs2 -n -R "stats" /dev/sdb1 | grep -i "Cluster Size" Block Size Bits: 12 Cluster Size Bits: 15 [root at sai93 ~]# cat /etc/redhat-release Red Hat Enterprise Linux Server release
2014 Aug 22
2
ocfs2 problem on ctdb cluster
Ubuntu 14.04, drbd Hi On a drbd Primary node, when attempting to mount our cluster partition: sudo mount -t ocfs2 /dev/drbd1 /cluster we get: mount.ocfs2: Unable to access cluster service while trying to join the group We then call: sudo dpkg-reconfigure ocfs2-tools Setting cluster stack "o2cb": OK Starting O2CB cluster ocfs2: OK And all is well: Aug 22 13:48:23 uc1 kernel: [
2008 Aug 21
5
VM node won't talk to host
I am trying to mount the same partition from a KVM ubuntu 8.04.1 virtual machine and on an ubuntu 8.04.1 host server. I am able to mount the partition just on fine on two ubuntu host servers, they both talk to each other. The logs on both servers show the other machine mounting and unmounting the drive. However, when I mount the drive in the KVM VM I get no communication to the host
2006 Mar 23
4
initial cluster setup
Hello list! Just pulled down ocfs2 this morning and installed it on two RHEL 4 2.6.9-34 systems. I have followed the manual, but am not seeing some things it describes and am not sure what the problem is. I guess my first question is - do I need to have a partition formatted as ocfs2 on all nodes in the cluster for this to work? I have formatted a partition via ocfs2console on node1, and
2008 Mar 31
4
SuSe Hangs when /etc/init.d/o2cb online
Hello, I have a DELL MD3000i, and a couple of servers that I want to connect to the array. I have setup the main server with CentOS 5. - 2.6.18-53.el5 x86_64 - ocfs2-tools-1.2.7-1.el5 - ocfs2console-1.2.7-1.el5 - ocfs2-2.6.18-53.el5-1.2.8-2.el5 # Kernel Module And two OpenSuSe 10.3 Servers - 2.6.22.5-31-default x86_64 - ocfs2-tools-1.2.6-18 - ocfs2console-1.2.6-18 The module has been already
2009 Oct 20
1
Fencing OCFS
I people, i install the ocfs in my virtual machine, with centos 5.3 and xen. But when i turn off the machine1, the ocfs start the fencing off the machine2. I read the doc in oracle.com, but could not solve the problem. Someone help? My conf and my package version. cluster.conf node: ip_port = 7777 ip_address = 192.168.1.35 number = 0 name = x1 cluster
2009 Jan 14
1
Transport endpoint is not connected while mounting....
Does anyone have any idea what to try next? Here are the steps I have taken and the problem: (I wanted to post my question on the first line before I explained the problem and what I have tried) ---------- Node 0 has the file system mounted just fine and works great. When trying to mount on Node 1: `mount.ocfs2 /dev/mapper/data /cluster/ data` I get this error after about 30 seconds:
2010 Sep 03
1
Servers reboot - may be OCFS2 related
Hello. What we have: 2x Debian 5.0 x64 - 2.6.32-20~bpo50+1 from backports DRBD + OCFS2 1.4.1-1 I have both node reboot every day on my tests. On heavy load it have to 1-3 hour to reboot, on idle about 20 hours. I not sure what it is a OCFS2 related but if I not mount OCFS2 partition - I don`t get this reboot. In attach screenshoot with system console on what error. Nothing special in logs.