Displaying 20 results from an estimated 6000 matches similar to: "OCFS2 v1.4 hangs"
2009 Jun 10
6
mount.ocfs2: Transport endpoint is not connected while mounting
Hi All,
I have a cluster of four nodes and one will not join the cluster. I have
two IPs on each node, one external and one internal. I have tried
changing around the IPs in the /etc/ocfs2/cluster.conf and that helped -
at least I recovered three of the machines. Any suggestions on where
else to look?
Best Regards
John
(23359,2):o2net_connect_expired:1637 ERROR: no connection established
with
2009 Jul 28
1
LUN Aggregation
Greetings.
Is there an approved way to aggregate LUNs when using OCFS2? I have several 1TB LUNs I'd
like to make into a single filesystem. Would I use something like Linux software raid?
Brett
2008 Dec 18
1
Filesystem Block Size w// DB_BLOCK_SIZE
Hello All,
We're hosting DB1 and DB2 with db_block_size set to 8K, 16K respectively
File system creation is done with mkfs.ocfs2 -b 4K -C 32K -N 4 -L LABLE
/dev/mapper/xxxx
Mount is done with: ocfs2 _netdev,datavolume,nointr 0 0
I'd like to know if we can separate most of the tablespaces on different
LUNs, even if they're on the same disk group sometimes, is it
2009 Aug 17
1
NIC Bonding
Hi,
I have 2 NIC and wants to use both NIC for load balancing and fault
tolerance.
- Quick question if NIC bonding in Redhat is supported by OCFS2?
- If yes which mode bonding is supported?
Thanks
Raheel
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
2010 Jul 14
2
some beginner questions
Hi,
I'am new to ocfs2 filesystem and I have some questions about it.
I installed three server according to the user guide from
http://oss.oracle.com/projects/ocfs2/dist/documentation/v1.4/ocfs2-1_4-usersguide.pdf
For every single server I have a working ocfs2 partition
/dev/sdb1 on /mnt/oc1 type ocfs2 (rw,_netdev,heartbeat=local)
As I understand the ocfs2 system I can use now these
2009 Dec 15
2
Subversion server: v1.4 (centos) vs. v1.6 (rpmforge)
Hi,
I'm planning to upgrade an old public/internal development
infrastructure and will use CentOS 5.4 x86_64 as basis.
One of the critical server is Subversion (as an Apache httpd module).
We currently use Subversion v1.4 on the server and v1.6 on our clients.
While I was very happy with the v1.4 server over the years, we now
sometimes have weird issues which seem related to compatibility
2007 Jun 29
2
v1.4.x ready yet?
Hi All,
Eagerly waiting for v1.4.x to mature a bit before getting serious about it.
Is it ready for production yet? If that's too general, where is it in terms
of stability compared to where 1.2.x is now. Anyone running it successfully
in production environment and if so what sort of config do you have?
2010 Nov 02
1
disabling self-fencing
Hi all,
on my nodes I have another cluster manager running that takes care of
fencing the node, is it safe to disable ocfs2 self-fencing?
Details:
I'm using ocfs2 over drbd in master/master aka primary/primary mode.
In case of loss of network connectivity I would like to disconnect the
drbd device, invalidate it and unmount the filesystem but ocfs2
reboots the node....
Is it possible to
2006 Jul 28
3
Private Interconnect and self fencing
I have an OCFS2 filesystem on a coraid AOE device.
It mounts fine, but with heavy I/O the server self fences claiming a
write timeout:
(16,2):o2hb_write_timeout:164 ERROR: Heartbeat write timeout to device
etherd/e0.1p1 after 12000 milliseconds
(16,2):o2hb_stop_all_regions:1789 ERROR: stopping heartbeat on all
active regions.
Kernel panic - not syncing: ocfs2 is very sorry to be fencing this
2008 Mar 05
3
cluster with 2 nodes - heartbeat problem fencing
Hi to all, this is My first time on this mailinglist.
I have a problem with Ocfs2 on Debian etch 4.0
I'd like when a node go down or freeze without unmount the ocfs2 partition
the heartbeat not fence the server that work well ( kernel panic ).
I'd like disable or heartbeat or fencing. So we can work also with only 1
node.
Thanks
2010 May 26
1
Failover testing problem and a heartbeat question
We have a setup with 15 hosts fibre attached via a switch to a common SAN. Each host has a single fibre port, the SAN has two controllers each with two ports. The SAN is exposing four OCFS2 v1.4.2 volumes. While performing a failover test, we observed 8 hosts fence and 2 reboot _without_ fencing. The OCFS2 FAQ recommends a default disk heartbeat of 31 - 61 loops for multipath io users. Our initial
2007 Feb 06
2
Network 10 sec timeout setting?
Hello!
Hey didnt a setting for the 10 second network timeout get into the
2.6.20 kernel?
if so how do we set this?
I am getting
OCFS2 1.3.3
(2201,0):o2net_connect_expired:1547 ERROR: no connection established
with node 1 after 10.0 seconds, giving up and returning errors.
(2458,0):dlm_request_join:802 ERROR: status = -107
(2458,0):dlm_try_to_join_domain:950 ERROR: status = -107
2010 Dec 09
1
[PATCH] Call userspace script when self-fencing
Hi,
According to comments in file fs/ocfs2/cluster/quorum.c:70 about the
self-fencing operations :
/* It should instead flip the file
?* system RO and call some userspace script. */
So, I tried to add it (but i did'nt find a way to flip the fs in RO).
Here is a proposal for this functionnality, based on ocfs2-1.4.7.
This patch add an entry 'fence_cmd' in /sys to specify an
2008 Jul 01
5
ocfs2 fencing problem
Hi, Sunil or Tao,
I have a 4 nodes OCFS2 cluster running OCFS2 1.2.8 on SuSE 9 SP4. When I
tried to do failover testing (shutting down one node), the whole cluster
hung (I can not even login to any server in the cluster). I have to
bring all of them up and then be able to use the system. What kind of
behavior is it? Is it the fence of OCFS2? Below is my configuration.
aopcer13:~ #
2007 Apr 27
1
has anyone experienced problems with ocfs2 1.2.5-1 using Emulex LP10000 HBA cards and EMC CX700 SAN's?
Does anyone have any experience with Emulex HBA cards (LP10000) using OCFS2,
Linux AS4 U4 x86_64 AMD? I'm trying to find out whether this is a verified
combination, if anyone has successfully used it. I have that
hardware/sofware combination, and am experiencing
stability/performance/panic/hang issues with OCFS2.
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
2007 Jul 29
1
6 node cluster with unexplained reboots
We just installed a new cluster with 6 HP DL380g5, dual single port Qlogic 24xx HBAs connected via two HP 4/16 Storageworks switches to a 3Par S400. We are using the 3Par recommended config for the Qlogic driver and device-mapper-multipath giving us 4 paths to the SAN. We do see some SCSI errors where DM-MP is failing a path after get a 0x2000 error from the SAN controller, but the path gets puts
2007 Nov 29
1
Troubles with two node
Hi all,
I'm running OCFS2 on two system with OpenSUSE 10.2 connected on fibre
channel with a shared storage (HP MSA1500 + HP PROLIANT MSA20).
The cluster has two node (web-ha1 and web-ha2), sometimes (1 or 2 times
on a month) the OCFS2 stop to work on both system. On the first node I'm
getting no error in log files and after a forced shoutdown of the first
node on the second I can see
2010 Jan 18
1
Getting Closer (was: Fencing options)
One more follow on,
The combination of kernel.panic=60 and kernel.printk=7 4 1 7 seems to
have netted the culrptit:
E01-netconsole.log:Jan 18 09:45:10 E01 (10,0):o2hb_write_timeout:137
ERROR: Heartbeat write timeout to device dm-12 after 60000
milliseconds
E01-netconsole.log:Jan 18 09:45:10 E01
(10,0):o2hb_stop_all_regions:1517 ERROR: stopping heartbeat on all
active regions.
2009 Nov 17
1
[PATCH 1/1] ocfs2/cluster: Make fence method configurable
By default, o2cb fences the box by calling emergency_restart(). While this
scheme works well in production, it comes in the way during testing as it
does not let the tester take stack/core dumps for analysis.
This patch allows user to dynamically change the fence method to panic() by:
# echo "panic" > /sys/kernel/config/cluster/<clustername>/fence_method
Signed-off-by: Sunil
2008 Aug 21
1
Shared Storage Options
Hello all.
I would like to canvas some opinions on options for shared storage in a Xen cluster. So
far I've experimented with using iSCSI and clvm which mixed success.
The primary concern I have with both of these options is that there seems to be no obvious
way to ensure exclusive access to the LUN/device to the VM I want to run. On a couple of
occasions during my playing I've